2 Rounds of Client Interview
Must have Strong SQL and Python and Spark Data Pipelines and ETL Framework
- Develop and automate large scale, high-performance data processing systems (batch and/or streaming) to drive Airbnb business growth and improve the product experience.
- Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
- 8+ years of relevant industry experience
- Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions
- Working knowledge of relational databases and query authoring (SQL).
- Good communication skills, both written and verbal
- Strong experience using ETL framework (ex: Airflow, Flume, Oozie etc.) to build and deploy production-quality ETL pipelines.
- Experience building batch data pipelines in Spark Scala.
- Strong understanding of distributed storage and compute (S3, Hive, Spark)
- General software engineering skills (Java or Python, Github)