Company: J.P. Morgan
Answer these behavioral prompts:
1) Tell me about yourself and a recent project you’re proud of.
2) Describe a time failing test cases were blocking delivery near a deadline—what actions did you take, what trade-offs did you consider, and what was the outcome?
3) Describe a disagreement with a teammate and how you resolved it.
4) Share a significant failure, how you handled it, and what you learned.
5) Why JPMorgan and this team?
Quick Answer: This question evaluates behavioral and leadership competencies—communication, conflict resolution, accountability, decision-making under pressure, and the ability to convey impact through structured examples about projects, failures, and trade-offs.
Solution
# How to approach behavioral answers
- Use STAR for each story: Situation → Task → Action → Result (+ Reflection/Learning).
- Lead with outcomes and numbers (latency, throughput, defect rate, SLA, cost, time saved).
- Show judgment: trade-offs, risk, stakeholders, and constraints (security, compliance, reliability).
- Keep answers 60–120 seconds each; go deeper if probed.
- The example stories below are illustrative—replace domain or tech to match your background.
## 1) Tell me about yourself and a recent project you’re proud of
Structure:
- Present: What you do now and your focus/strengths.
- Past: 1–2 relevant experiences showing impact at scale.
- Future: Why this role/team fits your trajectory.
- Project: One concise story with measurable impact.
Sample answer:
- Present: I’m a software engineer focused on backend systems and reliability. I build and scale event-driven services in Java/Kotlin with Kafka, and I care a lot about observability and testing.
- Past: Over the last 3 years, I’ve delivered services processing 50k+ msgs/sec, reduced p95 latency from 280 ms to 120 ms, and cut incident volume by 35% by adding idempotency and circuit breakers.
- Project I’m proud of: Recently, I led a real-time risk alerting service. Situation: Compliance needed alerts under 1 second to meet a new SLA. Task: Deliver a low-latency, highly available pipeline before quarter-end. Actions: I designed a Kafka-based stream with a stateful processor (Flink), added a Redis cache for hot reference data, and introduced feature flags and shadow traffic to de-risk rollout. I built SLOs and dashboards (p95 latency, error budget) and coordinated load tests to 15k events/sec. Result: We hit 750 ms p95, 99.99% availability, and reduced false positives by 22% via better deduplication. The project went live a week early and has had zero Sev-1s in six months. Future: I’m looking to apply this systems and reliability mindset to mission-critical services at JPMorgan where scale, correctness, and security matter.
Tips:
- If you’re earlier career, emphasize internships and class projects with similar structure and smaller numbers.
- Keep the “about me” to ~45 seconds; spend the rest on the project.
## 2) Failing test cases blocking delivery near a deadline
What interviewers seek: Triage skills, risk management, communication, and quality bar under time pressure.
Sample STAR answer:
- Situation: Two days before a quarterly release, 14 integration tests started failing in CI after merging a dependency upgrade and a schema change.
- Task: Decide whether we can ship on time without compromising reliability or compliance.
- Actions:
- Triage: I categorized failures into P1 (correctness/security), P2 (non-critical behavior), P3 (flaky/environmental). I reproduced locally with the same container images used in CI to eliminate “works on my machine.”
- Contain blast radius: Introduced a feature flag to gate the new schema usage; toggled to off in prod paths while keeping the deployment train moving.
- Parallelize fixes: Paired with QA to update test data for new nullability rules; stubbed an external service that had changed rate limits; reverted only the risky transitive dependency via a temporary resolution override.
- Communicate: Sent a risk update with options: (1) ship with flag off and track two P2 issues, (2) slip release by 48 hours to fully validate. I aligned with PM and compliance on acceptable risk.
- Guardrails: Added a canary stage requiring P1 green + error budget check; tightened merge rules to block on flaky test tags until deflaked.
- Trade-offs considered:
- Scope vs. schedule: Limited scope by flagging off the schema change to avoid data migration risk while meeting the date.
- Short-term override vs. long-term stability: Temporary dependency pin to ship; ticketed follow-up to safely adopt the upgrade with contract tests.
- Test completeness vs. operational validation: Accepted P2 UI mismatch for non-critical admin screen; compensated with increased telemetry during canary.
- Result: Shipped on time with zero Sev-1s. Closed all P1 failures; tracked two P2 items and fixed within a week. CI flakiness dropped 60% after deflaking and adding contract tests.
Pitfalls to avoid:
- Don’t say “we disabled tests and shipped.” Explain risk assessment, flags, canaries, and rollback plans.
- Always show stakeholder alignment and documented follow-ups.
## 3) Disagreement with a teammate and resolution
What interviewers seek: Collaboration, humility, evidence-driven decisions.
Sample STAR answer:
- Situation: We needed to add a new pricing recommendation capability. I proposed extending our existing service with a new module; a teammate preferred a new microservice for clean boundaries.
- Task: Choose an approach balancing latency, ownership, and operational complexity.
- Actions:
- Clarified constraints: 100 ms p95 latency budget and shared auth; team had limited on-call bandwidth.
- Proposed a data-driven spike: We built two prototypes in 2 days—one embedded module using Postgres JSONB for config and one separate service with gRPC.
- Measured: The embedded module added 12 ms p95; the separate service added 48 ms p95 plus an extra deployment and pager rotation.
- Facilitated decision: We reviewed results with the team, explicitly mapping trade-offs (latency, reliability, ownership, blast radius). We agreed on a hybrid: embedded for read path, small sidecar process for offline training jobs to keep boundaries clean.
- Documented in an ADR and planned a revisit in 3 months when traffic scaled.
- Result: Met latency goals, avoided new on-call burden, and preserved a path to split later. The teammate later led the split when QPS tripled; our earlier ADR made it straightforward.
Principles to highlight:
- Seek shared goals first, not “winning.”
- Use small experiments and ADRs to depersonalize choices.
- Time-box spikes; agree on revisit criteria.
## 4) Significant failure, response, and learning
What interviewers seek: Ownership, transparency, and systemic fixes.
Sample STAR answer:
- Situation: I introduced a memory leak by caching large protobuf objects without TTL in a long-lived worker.
- Task: Stop the incident, communicate, and prevent recurrence.
- Actions:
- Owned it: Declared a Sev-2, joined incident command, wrote the comms updates.
- Contained: Rolled back the release, drained the queue, and temporarily reduced batch size to relieve pressure. Added a short-term JVM heap bump to stabilize while we fixed the leak.
- Found root cause: Heap dump + MAT showed unbounded cache growth keyed by request IDs. We added size-based eviction and TTL, switched to caching parsed feature vectors instead of full objects, and introduced load tests with heap assertions.
- Systemic fixes: Added canary with synthetic traffic targeting the cache path; created an automated regression test that fails on retained-size deltas beyond threshold; documented a performance review checklist for memory-sensitive changes.
- Result: MTTR was 42 minutes; no data loss due to idempotent processing. Post-fix, memory headroom improved by 30%, and we’ve had zero repeats in 9 months.
- Learning: Optimize for safety-first rollbacks, make the safe thing the easy thing (canaries, guardrails), and treat failures as team process improvements, not individual blame.
Tips:
- Be specific about the flaw, the fix, and the prevention.
- Show how you protected customers and what changed in your process.
## 5) Why JPMorgan and this team?
Structure your answer around:
- Impact at scale: Mission-critical systems, large data volumes, strict reliability/security.
- Engineering challenges: Low latency, resiliency, observability, compliance-by-design.
- Team fit: How your experience maps to their stack/problem space.
- Growth: Learning, mentorship, and cross-team collaboration.
Sample answer:
- I’m drawn to JPMorgan because it operates at a scale where reliability, performance, and security aren’t optional—they’re core to customer trust. I enjoy building systems with clear SLOs, strong observability, and rigorous change management, which aligns with the engineering culture here.
- This team’s focus on [payments/risk/markets/platform] matches my background in event-driven services and resilience. For example, I’ve shipped Kafka-based pipelines with exactly-once semantics, added contract tests to stabilize integrations, and used feature flags and canaries to reduce change failure rate. I’m excited to apply those practices to high-volume, low-latency workloads where milliseconds and correctness truly matter.
- I’m also motivated by the opportunity to learn from deep domain experts, contribute to internal platforms and tooling, and grow into leading projects that raise availability and developer productivity.
Customization prompts for you:
- Name 2–3 concrete team challenges you’re excited about (e.g., sub-100 ms end-to-end latency, zero-downtime schema migration, controls & auditability).
- Map your experience with numbers to those challenges.
- Mention 1–2 ways you’ll add value in your first 90 days (e.g., improve CI flakiness, instrument SLOs, reduce on-call toil).
# Final guardrails
- Keep customer and confidential data out of your stories; anonymize vendors if needed.
- Use real, verifiable metrics you can defend.
- Balance ownership with collaboration—credit the team, but be clear about your actions.
- Have 1–2 backup examples for each prompt in case interviewers probe for variety.