Roku is looking to build a world-class big data platform that allows both internal and external stakeholders to use data for business growth. The Senior Data Engineer will be responsible for designing data models and creating scalable data pipelines to collect business metrics across all Roku devices.
Requirements
- 5+ years of professional experience as a data or software engineer
- 3+ years of hands-on experience with SQL is required
- Proficiency in at least one scripting language, Python, is required
- Experience with big data technologies such as HDFS, YARN, MapReduce, Hive, Kafka, Spark, Airflow, and Presto
- Proficiency in data modeling, including the design, implementation, and optimization of conceptual, logical, and physical data models to support scalable and efficient data architectures
- Experience with cloud AWS/GCP and Looker is a plus
Responsibilities
- Build highly scalable, available, and fault-tolerant distributed data processing systems for both batch and streaming, processing tens of terabytes of data ingested daily, as well as managing a petabyte-sized data warehouse
- Develop robust data solutions and streamline diverse datasets into simplified models to promote self-service
- Develop data pipelines that ensure high quality and are resilient to low-quality data sources
- Take responsibility for data mapping, business logic, transformations, and data quality
- Debugging low-level systems, measuring performance, and optimizing large production clusters
- Engage in architectural discussions, influence the product roadmap, and take ownership of new projects
- Support and maintain existing platforms while transitioning to newer technology stacks and architectures
Other
- A Bachelor’s or Master’s of Science in Computer Science is preferred
- Collaborate with cross-functional teams such as developers, analysts, and operations to execute deliverables