PracHub
QuestionsPremiumLearningGuidesCheatsheetNEWCoaches
|Home/Behavioral & Leadership/Bank of America

Showcase initiative and collaboration examples

Last updated: Mar 29, 2026

Quick Overview

This prompt evaluates initiative, ownership, quantitative rigor, complex problem-solving, and relationship-building within a data scientist role. Commonly asked to gauge leadership, cross-functional impact, and measurable technical outcomes, it is categorized as Behavioral & Leadership in the data science domain and emphasizes practical application of analytical methods alongside conceptual reasoning about trade-offs and collaboration.

  • medium
  • Bank of America
  • Behavioral & Leadership
  • Data Scientist

Showcase initiative and collaboration examples

Company: Bank of America

Role: Data Scientist

Category: Behavioral & Leadership

Difficulty: medium

Interview Round: Take-home Project

##### Question Which one of your key accomplishments best illustrates your personal initiative and willingness to push beyond what is required? Walk through one of your favorite quantitative or technical projects that you completed at school or with a previous employer. What was the end goal? What technical tools did you use to solve the problem? Why did you choose these tools? What was the outcome? Tell us about a time you solved a complex problem that required a lot of thought and careful analysis on your part. In your response, please describe the problem, the analysis you performed, your solution and why you chose it, obstacles you had to overcome, and how your solution was implemented. Describe a time when you actively attempted to develop a strong relationship with a teammate, manager or customer/client. In your response, please share the specific actions you took to build the relationship, any challenges you faced, how you addressed them, and what resulted.

Quick Answer: This prompt evaluates initiative, ownership, quantitative rigor, complex problem-solving, and relationship-building within a data scientist role. Commonly asked to gauge leadership, cross-functional impact, and measurable technical outcomes, it is categorized as Behavioral & Leadership in the data science domain and emphasizes practical application of analytical methods alongside conceptual reasoning about trade-offs and collaboration.

Solution

# How to approach behavioral/technical responses - Use STAR: Situation, Task, Action, Result. Add Reflection when appropriate. - Quantify impact (%, $, time saved, users affected). Tie model metrics to business outcomes. - Surface trade-offs, constraints, and why you chose one approach over alternatives. - Add guardrails/validation for any experiments or models (SRM checks, leakage checks, drift monitoring). --- ## 1) Initiative and Ownership — Model Answer Situation: Product teams were running many A/B tests, but 40–50% were underpowered, causing inconclusive results and wasted cycles. Task: Though not asked to, I wanted to reduce inconclusive tests by enabling PMs/analysts to self-serve power and sample size planning. Action: - Built a self-serve experiment design app (Streamlit + Python) with calculators for proportions and means, pre/post CUPED variance reduction, and minimum detectable effect (MDE) guidance. - Implemented sample size for proportions: n = 2 * (Z_{1−α/2} + Z_{1−β})^2 * p(1−p) / Δ^2, with sensible priors for baseline p and business-relevant Δ. - Added guardrails: SRM chi-square check, sequential-look guidance, and a checklist (primary metric, segmentation, pre-registration, power). - Ran 3 workshops and created a 2-page playbook with examples. Result: - Share of adequately powered tests increased from 52% to 86% over 2 quarters. - Average test duration decreased by 18% due to CUPED and better MDE setting. - Estimated 25–30 analyst hours/month saved; clearer test decisions improved trust in experimentation. Why it shows initiative: I identified a cross-team pain point, built a practical tool, trained users, and measured impact without being asked. --- ## 2) Favorite Quantitative/Technical Project — Model Answer End goal: Reduce false positives (legitimate transactions incorrectly flagged) in a fraud screening system while maintaining fraud recall. Tools and why: - SQL (Snowflake) for feature extraction and time-based joins — scalable warehouse, strong window functions. - Python (pandas, scikit-learn, LightGBM) — tabular data; gradient boosting excels with heterogeneous features and missingness. - MLflow for experiment tracking; SHAP for explainability; Great Expectations for data quality checks; Airflow for batch scoring. Methods: - Label definition: Used only pre-authorization features to avoid leakage from post-transaction outcomes. - Feature engineering: Rolling aggregates (txn count/amount by card, device, merchant over 1h/24h/7d), velocity features, device-merchant affinity, geodistance. - Validation: Time-based splits (k-fold by month) to respect temporal dependence; class-weighted loss to address imbalance. - Baseline vs. model: Logistic regression baseline, then LightGBM with Bayesian optimization. Calibrated probabilities (isotonic) and tuned threshold for asymmetric costs. - Objective: Minimize expected cost: E[cost] = c_FP * FP + c_FN * FN with constraint recall ≥ target. Small numeric example: - Sample month: 100,000 txns, 200 fraud (0.2%). Baseline model achieves recall 95% with FP rate 0.50%. - Baseline FP = 0.005 × 99,800 ≈ 499; FN = 10. - New LightGBM tuned to same recall (95%) yields FP rate 0.36%. - New FP ≈ 359; FN = 10. - If c_FP = $5 (customer friction) and c_FN = $500 (loss), baseline cost ≈ 499×5 + 10×500 = $7,495; new cost ≈ 359×5 + 10×500 = $6,795 → ~$700/month-sample savings; scaled to production volume this was >$1.2M/year. Obstacles and mitigations: - Data leakage: Excluded features derived after authorization; enforced time-aware pipelines. - Class imbalance: Class weights and focal loss experiments; monitored PR-AUC, recall at fixed FPR. - Drift: Implemented PSI/KS monitoring on key features and calibration drift checks; set alerts for retraining. Implementation: - Deployed a batch scoring DAG (Airflow) with canary routing (10% traffic) and a kill switch. - Explanations: SHAP top features and reason codes logged for each decision for risk review. - A/B validation: Guardrails included SRM checks, pre-registered metrics (recall at fixed FPR, customer contact rate), and a 2-week holdout. Outcome: - Reduced false positives by 28% at constant fraud recall (95%). - Customer contact rate decreased by 22%; improved satisfaction (CSAT +3.1 pts) and analyst review load −18%. Why these tools: Tree-based boosting handled sparse/categorical interactions and nonlinearity; SHAP provided transparency needed by risk; MLflow and Great Expectations ensured reproducibility and data quality. Pitfalls to avoid: - Evaluating on randomly split data causing temporal leakage. - Using ROC-AUC alone; for rare events, focus on PR-AUC and cost-sensitive metrics. - Deploying without calibration can harm thresholding; always check calibration curves. --- ## 3) Complex Problem Solving — Model Answer Problem: An onboarding UI change “won” an A/B test on conversion (+2.4 pts), yet downstream revenue and activations didn’t move. Stakeholders questioned measurement integrity. Analysis: - Suspected Simpson’s paradox due to channel mix shift. Segmented by acquisition channel (Paid, Organic) and device. - Within-stratum effects were ~0, but treatment had disproportionately more Paid traffic (higher baseline conversion), inflating the aggregate. Illustrative numbers: - Control: Paid 30k @ 8% (2,400), Organic 70k @ 2% (1,400) → overall 3.8% (3,800/100k). - Treatment: Paid 70k @ 8% (5,600), Organic 30k @ 2% (600) → overall 6.2% (6,200/100k). - Within each stratum, treatment effect = 0; the “lift” came from mix shift, not the UI. Solution and rationale: - Reanalyzed with stratified estimates (Cochran–Mantel–Haenszel weighting) matching historical channel proportions → net effect ≈ 0. - Implemented stratified randomization in our experimentation platform and added reweighting in the analysis pipeline. - Pre-registered segmentation and added SRM checks (chi-square p < 0.01 alert). Obstacles: - Some strata had low sample sizes. We merged adjacent device strata with similar baselines and used variance-stabilizing transforms; documented minimum stratum size rules. - Pushback from teams invested in the “win.” Addressed via a neutral postmortem with clear math and a rerun using stratified allocation. Implementation: - Platform change: randomization key included user_id × channel bucket to balance arms per channel. - Analysis templates added: stratified diff-in-proportions with delta-method CIs and reweighting utilities. Outcome: - Rerun showed 0.1 pt (ns) effect; we avoided shipping a non-impactful change affecting millions of users. - New guardrails reduced SRM incidents by 80% over the next quarter; leadership confidence in experimentation improved. Reflection: Complex problems often need you to check identification assumptions (no confounding, balanced allocation). Stratification and SRM guardrails are low-cost, high-impact fixes. --- ## 4) Relationship Building — Model Answer Situation: A senior Risk Manager was skeptical of ML-driven decisions due to explainability and compliance concerns, blocking model adoption. Actions: - Scheduled recurring 1:1s to understand their risk appetite, required disclosures, and audit trail needs. - Co-designed acceptance criteria: confusion matrix thresholds, reason-code coverage, fairness checks across protected groups. - Built a simple “risk lens” dashboard: for any threshold, show expected FP/FN counts, cost curves, and top SHAP reason codes. - Ran a pilot “champion–challenger” with a small exposure cap and daily review, enabling gradual trust-building. Challenges and how I addressed them: - Different vocabularies: I mapped model metrics to risk terms (e.g., recall as “coverage,” precision as “purity”) and provided glossary tooltips. - Time constraints: I prepared 1-page briefs before meetings and async Loom walkthroughs to reduce meeting load. Outcome: - Model approved for broader rollout with jointly owned thresholds; review time per case −30%; false positives −18% with no recall loss. - We established a repeatable template for future models, shortening approval cycles by ~25%. Reflection: Trust is built by aligning on incentives, making trade-offs explicit, and creating shared artifacts (dashboards, criteria) that reduce ambiguity. --- # Summary playbook you can reuse - Frame stories with STAR; quantify impact and tie it to business outcomes. - For models: prevent leakage, use time-aware validation, focus on cost-sensitive metrics, and provide explainability. - For experiments: plan power/MDE, check SRM, pre-register metrics/segments, and stratify if needed. - For relationships: align on shared KPIs, build transparency, and iterate with low-risk pilots.

Related Interview Questions

  • Describe initiative, project, analysis, and relationship-building - Bank of America (medium)
  • Describe building a professional relationship - Bank of America (medium)
  • Explain motivations, projects, accomplishments, and teamwork - Bank of America (hard)
  • Answer behavioral prompts for quant internship - Bank of America (medium)
Bank of America logo
Bank of America
Aug 4, 2025, 10:55 AM
Data Scientist
Take-home Project
Behavioral & Leadership
1
0

Behavioral & Technical Prompts (Data Scientist — Take‑Home Interview)

Provide concise, structured responses that demonstrate ownership, quantitative rigor, and cross-functional impact. When possible, use STAR (Situation, Task, Action, Result) and quantify outcomes.

1) Initiative and Ownership

Which one of your key accomplishments best illustrates your personal initiative and willingness to go beyond what was required?

  • Describe the situation and what was expected.
  • Explain what you did beyond the scope.
  • Share measurable results and organizational impact.

2) Favorite Quantitative/Technical Project

Walk through one of your favorite quantitative or technical projects completed at school or a previous employer.

  • End goal and business context.
  • Technical tools used and why you chose them.
  • Methods, models, or analyses performed.
  • Obstacles and how you addressed them (e.g., data quality, leakage, class imbalance).
  • Outcome, metrics moved, and how it was implemented.

3) Complex Problem Solving

Tell us about a time you solved a complex problem that required careful analysis.

  • Problem statement and constraints.
  • Analysis you performed; frameworks or methods used.
  • Your solution and rationale vs. alternatives.
  • Obstacles, trade-offs, and implementation details.
  • Results and what you learned.

4) Relationship Building

Describe a time you actively developed a strong relationship with a teammate, manager, or customer/client.

  • Specific actions you took to build trust and alignment.
  • Challenges you faced and how you addressed them.
  • The outcome and how it improved collaboration or results.

Solution

Show

Comments (0)

Sign in to leave a comment

Loading comments...

Browse More Questions

More Behavioral & Leadership•More Bank of America•More Data Scientist•Bank of America Data Scientist•Bank of America Behavioral & Leadership•Data Scientist Behavioral & Leadership
PracHub

Master your tech interviews with 7,500+ real questions from top companies.

Product

  • Questions
  • Learning Tracks
  • Interview Guides
  • Resources
  • Premium
  • For Universities
  • Student Access

Browse

  • By Company
  • By Role
  • By Category
  • Topic Hubs
  • SQL Questions
  • Compare Platforms
  • Discord Community

Support

  • support@prachub.com
  • (916) 541-4762

Legal

  • Privacy Policy
  • Terms of Service
  • About Us

© 2026 PracHub. All rights reserved.