The company is looking to build and grow a clinical data registry platform over many years to drive advancements in healthcare, improve patient outcomes, and shape the future of medicine by managing, optimizing, and expanding data infrastructure.
Requirements
- Advanced SQL and Python
- Expertise in the design and construction of Big Data Lakes and Data Warehouses capable of ingesting, standardizing, and serving billions of data rows spanning diverse datasets ranging from tens to hundreds
- Experience building dynamic, metadata driven pipelines and analyses
- Building and managing fully automated data pipelines (ETL, ELT, ELTL)
- Designing and building data interfaces to source systems
- Combining and transforming data into the appropriate format for storage
- Developing data sets for analytics purposes
Responsibilities
- Designing, building, and maintaining data infrastructure
- Managing, optimizing, and expanding our data infrastructure
- Building and managing fully automated data pipelines (ETL, ELT, ELTL)
- Designing and building data interfaces to source systems
- Combining and transforming data into the appropriate format for storage
- Developing data sets for analytics purposes
- Developing pipelines that can handle common issues/errors in a robust and automated way
Other
- At least 6 years of industry working experience
- Working in Agile and DevOps environments
- Data modelling – Kimball, Data Vault, Star/Snowflake schema, Query-first etc.
- Data visualisation in Power BI, Tableau, Qlik or similar
- Architecting Data Platforms - designing BI/MI/Analytics solutions using Big Data, Relational or Streaming technologies