Real interview questions asked at Freight Tiger. Practice the most frequently asked questions and land your next role.
Freight Tiger data engineering interviews test your ability across multiple domains. These questions are sourced from real Freight Tiger interview experiences and sorted by frequency. Practice the ones that matter most. This set leans toward the medium-difficulty band most real interviews actually live in (6 of 8). Recurring themes are partition, spark, and join — these patterns appear most often in real interviews and reward the deepest preparation. Many of these questions also surface at FedEx Dataworks, so the preparation transfers across companies. Average answer is around 1 minute of reading — plan roughly 1 hour to work through the full set thoughtfully.
This collection contains 8 curated questions: 0 easy, 6 medium, and 2 hard. The distribution skews toward harder problems, reflecting the depth expected in senior-level interviews.
The most frequently tested areas in this set are partition (6), spark (4), join (3), optimization (2), sql (2), and window (1). Focusing on these topics will give you the highest return on your preparation time.
Medium-difficulty questions form the bulk of real interviews — spend the most time here and practice explaining your reasoning out loud. Hard questions often appear in senior and staff-level rounds; attempt them after you're comfortable with the basics. For each question, try answering before revealing the solution. Use our AI Mock Interview to simulate real interview conditions and get instant feedback on your responses.
Architecturally, how do Job–Stage–Task boundaries in Spark's execution model impact cluster sizing, shuffle cost, and when would you deliberately collapse or split stages?
What are the key components of the Spark execution model (Job, Stage, Task)?
How many cities does each department operate in? List the top 3 departments in terms of the most number of cities. In case of a tie, order by dept_id.
List every combination of dept_name, employee_name, and city such that the employee belongs to the department and the same city in which the department is located.
Add a column to the Employees table that shows the name of the employee with the next higher employee_id.
Find the third-highest salary for each department.
Write a PySpark job to find the top 3 employees of each department, where Age < 30 and Salary > department average salary.
Write a PySpark script to read a CSV file, filter rows where the age column is less than 18, and write the result to a new CSV file.
Get full access to 1,800+ expert answers, AI mock interviews, and personalized progress tracking.