Attain is looking to build and scale its data infrastructure to ensure data integrity and meaning, enabling high-impact decisions and products. They aim to create a more equitable and efficient data sharing ecosystem where consumers can access financial services and businesses can leverage data for better outcomes.
Requirements
- Proficiency in Python and familiarity with data orchestration tools (e.g., Airflow, GCP Workflows)
- Experience working with relational databases and large-scale cloud data warehouses
- Exposure to cloud infrastructure (GCP, AWS, or Azure)
- Familiarity with gRPC, Protobuf, or GraphQL
- Experience in AdTech, MarTech, or high-volume data environments is a plus
- Experience proactively working with modern AI-enhanced development tools (e.g., Cursor, Windsurf, Cody, etc.) and a curiosity for emerging AI workflows
Responsibilities
- Design and maintain data pipelines and orchestration layers using modern tools (e.g., Airflow, GCP Workflows)
- Own data workflows across cloud warehouses (BigQuery, Snowflake) and transactional stores
- Develop monitoring, alerting, and testing for data health and anomalies
- Partner with product, analytics, and engineering teams to validate business logic and definitions
- Define and maintain a clear, documented source of truth for key datasets
- Own the semantic correctness of our data — not just its arrival
- Write thoughtful, quality code that is readable, testable and easily maintainable
Other
- Enjoy building reliable data systems that power high-impact decisions and products
- Care deeply about data quality, semantics, and business logic – not just pipelines
- Proactively identify when data is misleading, incomplete, or broken
- Communicate clearly across technical and non-technical audiences
- Embrace feedback and thrive in a collaborative, fast-paced environment