Capgemini is looking to hire a data engineer to help its clients, the world's leading organizations, unlock the value of technology and build a more sustainable, more inclusive world by accelerating their dual transition to a digital and sustainable world.
Requirements
- Strong proficiency in Databricks (including Spark-based data processing).
- Strong programming skills in Python
- Advanced knowledge of SQL for querying and data modeling.
- Familiarity with Azure cloud and ADF
- Understanding of ETL frameworks, data governance, and performance tuning.
- Knowledge of CI/CD practices and version control (Git).
- Exposure to BI tools (Power BI, Tableau) for data visualization
Responsibilities
- Design, develop, and maintain scalable data pipelines and ETL processes using Databricks.
- Design an develop Python scripts for data transformation, automation, and integration tasks.
- Develop and optimize SQL queries for data extraction, transformation, and loading.
- Ensure data integrity, security, and compliance with organizational standards.
- Participate in code reviews and contribute to best practices in data engineering
Other
- Day One Onsite - Hybrid
- Collaborate with data scientists, analysts, and business stakeholders
- 3-5 years of professional experience in data engineering or related roles.