Real interview questions asked at Bristol Myers Squibb. Practice the most frequently asked questions and land your next role.
Bristol Myers Squibb data engineering interviews test your ability across multiple domains. These questions are sourced from real Bristol Myers Squibb interview experiences and sorted by frequency. Practice the ones that matter most. This set leans toward the medium-difficulty band most real interviews actually live in (4 of 9). Recurring themes are partition, spark, and window — these patterns appear most often in real interviews and reward the deepest preparation. Many of these questions also surface at Wipro, so the preparation transfers across companies. Average answer is around 1 minute of reading — plan roughly 1 hour to work through the full set thoughtfully.
This collection contains 9 curated questions: 3 easy, 4 medium, and 2 hard. There's a strong foundation of fundamentals-focused questions — ideal for building confidence before tackling advanced topics.
The most frequently tested areas in this set are partition (4), spark (3), window (2), join (2), sql (2), and optimization (1). Focusing on these topics will give you the highest return on your preparation time.
Start with the easy questions to warm up and solidify fundamentals. Medium-difficulty questions form the bulk of real interviews — spend the most time here and practice explaining your reasoning out loud. Hard questions often appear in senior and staff-level rounds; attempt them after you're comfortable with the basics. For each question, try answering before revealing the solution. Use our AI Mock Interview to simulate real interview conditions and get instant feedback on your responses.
Write a query to find the top three highest-paid employees in each department using window functions.
How do you see your career evolving in the next 3-5 years?
If your team disagrees on the approach to solving a problem, how do you manage the situation?
Explain the architectural trade-offs when optimizing a query on 100M+ rows: indexing vs. partitioning vs. materialized views. When does each approach become cost-prohibitive or operationally burdensome, and how do you quantify impact?
How would you handle nulls in a SQL join? Provide examples using COALESCE.
What are your expectations for the role beyond the salary?
What is the most common performance bottleneck in Spark jobs, and how would you resolve it?
Write PySpark code to filter records based on specific conditions and add a calculated column.
Write a PySpark script to filter out invalid records from a dataset and calculate the average for a specific column, ensuring the schema is strictly defined at runtime.
Get full access to 1,800+ expert answers, AI mock interviews, and personalized progress tracking.