Establish a world-class Data Engineering team to support data-centric aspirations and promote well-informed decision-making.
Requirements
- Experience with programming languages like Python, Spark, and SQL.
- Experience managing large-scale data sets, handling Terabytes of data and billions of records (across structured and unstructured datasets)
- Experience implementing job orchestrations using tools like Databricks, Airflow, Talend, and others.
- Good experience using GCP services like Google Storage, Looker, Google Analytics, and so on
- Proficiency in CI/CD tools (Jenkins, GitHub action, or similar)
- Comfortable using BI tools like Tableau, Looker, or Superset
Responsibilities
- Collaborate with the data team by prioritizing a service-oriented approach and quick response times.
- Observe high data quality standards, measurement certification methods, and engineering best practices.
- Improve reporting workflows and efficiency, harnessing the power of automation whenever possible.
- Enhance the performance, reliability, and scalability of storage and compute layers of the data lake.
- Build scalable, resilient, and observable distributed systems processes.
- Master the different data tools, clouds (AWS, GCP), and data engineering practices in the CDI ecosystem.
Other
- 2+ years of hands-on experience in data engineering or software development.
- Bachelor's degree in Computer Science, Information Systems, or a related field.
- Flexible time off policies allowing you to take the time you need to be your whole self.
- Generous medical, dental, vision, STD, LTD, and life insurance
- 401(k) plan, with employer match
- Employer paid commuter benefit
- Support program for new parents
- Pet insurance and some of our offices are pet friendly!