Real interview questions asked at Citi. Practice the most frequently asked questions and land your next role.
Citi data engineering interviews test your ability across multiple domains. These questions are sourced from real Citi interview experiences and sorted by frequency. Practice the ones that matter most.
What is the difference between repartition and coalesce in Apache Spark?
What is the difference between SparkSession and SparkContext in Spark?
What is the difference between partitioning and bucketing in Spark, and when would you use bucketing?
What strategies can you use to handle skewed data in Spark?
What is the difference between Managed and External tables in Hive/Spark?
What is a window function? Explain with an example.
Explain the concept of checkpointing in Spark and why it is important.
Agile methodologies used?
An existing job running longer suddenly: how to analyze the issue?
How is Oozie called?
Oozie workflow files (how many used)?
Shell commands for renaming a file?
Shell: change permissions?
Shell: command to check processes running in the background?
Using shell, how to find the difference between two files?
What type of wrapper is used, or which language is used?
Amazon Deequ usage and what sort of quality checks are done using it?
Given 1TB of a file, how to check word count?
Shell: how to run jobs/scripts in the background?
How to view Oozie jobs?
Oozie join condition?
Partitioning a table with card details and transactions?
Teradata to Hadoop migration and handling data with SCD Type 2?
What is a Kafka topic, and how do you choose the number of partitions for it?
What is the role of a partition in Kafka, and how does it impact scalability?
Describe how to pass data between tasks in Airflow using XComs.
Explain the concept of RDD, DataFrame, and Dataset in PySpark.
Explain the concept of consumer groups in Kafka. How do they affect message processing?
Explain the difference between TriggerDagRunOperator and ExternalTaskSensor in Airflow.
How do you ensure data quality and consistency across different stages of a data pipeline?
How do you handle failures in Airflow tasks, and what retry strategies can you use?
How do you optimize a join operation in Spark for large datasets?
How would you design a Kafka-based pipeline for processing streaming data in real-time?
Methods to avoid duplicates in PySpark/Scala?
Usage of UDFs?
What is a DAG in Apache Airflow, and how is it used for scheduling workflows?
Describe an end-to-end data pipeline project you worked on, highlighting your role and the technologies used.
Describe how Kafka ensures data durability and fault tolerance.
Introduce your recent project, explaining its goal, architecture, tools, and technologies.
Download the complete interview prep bundle with expert answers. Study offline, on your commute, anywhere.