At Labcorp, data is at the core of transforming healthcare and improving patient outcomes worldwide. As a Data Engineer, you’ll be responsible for designing, developing, and optimizing data solutions that empower analytics, reporting, and data science initiatives across the organization.
Requirements
- Strong Python development experience.
- Familiarity with Databricks, Spark, Hive, AWS EMR/S3, DataStage (or similar systems).
- Understanding of CI/CD concepts, modern build pipelines, and DevOps best practices.
- In-depth knowledge of Databricks, Hive, DataStage, and Oracle SQL.
- Proficiency in SQL querying and query tuning for performance optimization.
- Experience with CI/CD tools (e.g., Jenkins, GitHub Actions, GitLab CI).
Responsibilities
- Design, develop, and maintain scalable data pipelines and ETL processes to support enterprise data platforms.
- Leverage Python, Databricks, Spark, Hive, AWS EMR/S3, DataStage, and other modern tools to process, transform, and optimize large data sets.
- Ensure high performance and reliability of data systems through monitoring, query optimization, and tuning techniques.
- Collaborate with data scientists, analysts, and business stakeholders to deliver clean, reliable, and accessible data.
- Implement and support CI/CD pipelines for data engineering solutions, ensuring reproducibility and smooth deployment into production environments.
- Support ongoing migration and modernization initiatives in Labcorp’s cloud data ecosystem.
- Apply best practices in data governance, security, and compliance.
Other
- Previous Labcorp internship experience (required).
- Hands-on experience in data engineering or equivalent demonstrated academic/project work.
- Work within an Agile team environment, contributing to sprints, stand-ups, and iterative delivery cycles.
- Prior experience working in an Agile/Scrum environment.