This question evaluates understanding of sequential partial-observation evaluation, mask-value selection and reward optimization, augmentation strategies for robustness, and early-exit policy design for classifiers under progressively revealed inputs, testing competencies in model calibration, evaluation metrics, distribution-shift reasoning, and trade-off analysis in the ML System Design domain at both conceptual and practical application levels. It is commonly asked because it probes system-level thinking about information sufficiency, metric-driven trade-offs between accuracy and masked information, robustness to masking and augmentation, and the ability to design and calibrate stopping and confidence policies for streaming or cost-sensitive inference pipelines.
You have a trained CNN classifier for FashionMNIST. Each image is grayscale, shaped 1×28×28, and normalized (e.g., to [0, 1] or standard z-score). You also have an evaluation notebook where you can run batched inference across the test set.
Assume a row-wise reveal protocol: at step k (0 ≤ k ≤ 28), the top k rows are visible and the remaining rows are replaced by a constant mask fill value m. You will analyze how much information in the early rows suffices for accurate classification, choose a global mask value m to maximize a defined reward, propose a training augmentation policy to improve that reward, and design an early-exit policy when pixels are revealed sequentially at test time.
Login required