Develop and automate large scale, high-performance data processing systems (batch and/or streaming) to drive Airbnb business growth and improve the product experience.
Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
8+ years of relevant industry experience
Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions
Working knowledge of relational databases and query authoring (SQL).
Good communication skills, both written and verbal
Strong experience using ETL framework (ex: Airflow, Flume, Oozie etc.) to build and deploy production-quality ETL pipelines.
Experience building batch data pipelines in Spark Scala.
Strong understanding of distributed storage and compute (S3, Hive, Spark)
General software engineering skills (Java or Python,Github)