Capital Rx is looking to transform pharmacy benefits and healthcare navigation by building and maintaining reliable batch and streaming data pipelines, modeling data in Snowflake with dbt, and improving data quality and observability.
Requirements
- Strong SQL and Python; solid understanding of ETL/ELT and dimensional/data vault modeling
- Hands-on dbt experience (models, tests, documentation, deployments)
- Experience with orchestration (Dagster or Airflow) and scheduling best practices
- Familiarity with Spark/Databricks and cloud data warehouses (Snowflake preferred)
- CI/CD for data (Git-based workflows, environments), and basic IaC familiarity
- Data observability and testing (dbt tests, Great Expectations/Soda, lineage tools)
- Performance tuning in Snowflake (warehouse sizing, clustering, caching)
Responsibilities
- Implement and maintain batch/streaming pipelines with Dagster, Databricks/Spark, and dbt
- Build, test, and document Snowflake data models and business logic in dbt
- Apply and improve data quality, testing, observability, and lineage standards
- Collaborate with cross-functional partners to define data contracts and interfaces
- Contribute to Capital Rx’s modular data platform and client-specific data configurations
- Participate in design reviews; propose scalable, maintainable patterns
- Monitor pipeline health, troubleshoot incidents, and drive root-cause fixes
Other
- 5+ years in data engineering, software engineering, or related field
- Understanding of security, governance, and privacy for PII/PHI (HIPAA awareness)
- Effective communication and collaboration with technical and non-technical stakeholders
- Bonus if you have Exposure to healthcare data (claims, eligibility, provider directories)