Chess.com needs to process a massive amount of activity data for social feeds and product optimization, including games, puzzles, chats, and analysis, to serve its global community of 200M+ chess players.
Requirements
- Experience building and supporting data pipelines
- Experiencing running data workloads in the cloud in GCP
- Experience with Airflow, Flink & BigQuery or similar technology
- Experience with SQL and multiple databases (MySQL, MongoDB, etc)
- Experience deploying infrastructure as code through Terraform
Responsibilities
- Proactively collaborate with teammates to design and implement our world-class data platform.
- Develop and maintain dbt framework and API services and libraries that provide our core metrics, cohorts, and forecasts.
- Write and maintain reliable and efficient Airflow/Spark/Flink jobs to clean and enrich our data before it reaches our data warehouse.
- Use BigQuery and Airflow and dbt to perform data aggregations supporting insight generation.
- Protect our player's PII and help support data compliance across our data stack.
- Build and support tools for our internal teams to use to run experiments and access data.
- Support query design and query review and enable team members to share data.
Other
- Strong collaboration and communication skills working in a fully distributed team
- Sense of ownership and responsibility
- Chess player
- Lifelong learner
- fully remote