The Lakeflow team is looking to simplify the entire data lifecycle, from ingestion to generative AI and everything in-between, with a unified platform.
Requirements
- Knowledge of database systems
- Knowledge of storage systems
- Knowledge of distributed systems
- Knowledge of performance optimization
- Experience with Apache Spark
- Experience with Delta Live Tables (DLT)
- Experience with Materialized Views
Responsibilities
- Design and implement query compilation and optimization
- Implement distributed query execution and scheduling
- Develop vectorized engine execution
- Manage resource management
- Coordinate transaction coordination
- Design efficient storage structures (encoding, indexes)
- Implement automatic physical data optimization
Other
- PhD in databases or systems
- Motivated by delivering customer value and influence
- Annual performance bonus eligibility
- Equity eligibility
- Benefits listed above
- US government license for access to export-controlled technology or source code may be required