DoorDash is looking to build and maintain a scalable, reliable, and efficient data infrastructure, specifically a large-scale data lakehouse platform, to enable high-throughput analytics, ensure data quality, and comply with financial and privacy regulations.
Requirements
- Proficiency in at least one programming language: Scala, Java, or Python
- Experience with Big Data infrastructure development and management
- Familiarity with open-source data technologies such as Iceberg, Spark, Airflow, Kafka, Flink is a plus
- Experience in improving the scalability, efficiency, and stability of data platforms
Responsibilities
- Design, develop, and maintain a scalable data lakehouse platform to support high-volume analytics and data accessibility
- Scale the lakehouse infrastructure to handle hundreds of petabytes efficiently and reliably
- Integrate the data platform seamlessly with various compute platforms to accelerate adoption
- Develop solutions to enhance the performance, stability, and scalability of data systems
- Contribute to the evolution of DoorDash’s data architecture and roadmap
- Participate in code reviews, testing, and documentation to ensure high-quality deliverables
- Stay updated with emerging data technologies and industry best practices to continuously improve the platform
Other
- Minimum of 2+ years of professional experience in software engineering, with a solid understanding of computer science fundamentals
- Willingness to relocate or be based in the Bay Area, Seattle, or New York
- Collaborate with cross-functional teams to understand business needs and translate them into technical solutions