Adtalem is looking to solve the problem of building data solutions that power strategic and tactical business decisions and support Analytics and Artificial Intelligence operations by implementing a data platform, data pipelines, and data governance policies.
Requirements
- Hands-on experience working with real-time, unstructured, and synthetic data
- Expert knowledge on Python programming and SQL.
- Experience with cloud platforms (AWS, GCP, Azure) and their data services
- Experience working with Airflow as workflow management tools and build operators to connect, extract and ingest data as needed.
- Familiarity with synthetic data generation and unstructured data processing
- Experience in AI/ML data pipelines and frameworks
- Experience in Real Time Data ingestion using GCP PubSub, Kafka, Spark or similar.
Responsibilities
- Architect, develop, and optimize scalable data pipelines handling real-time, unstructured, and synthetic datasets
- Design, develop, deploy and support high performance data pipelines both inbound and outbound.
- Model data platform by applying the business logic and building objects in the semantic layer of the data platform.
- Leverage streaming technologies and cloud platforms to enable real-time data processing and analytics
- Optimize data pipelines for performance, scalability, and reliability.
- Implement CI/CD pipelines to ensure continuous deployment and delivery of our data products.
- Develop REST APIs to expose data to other teams within the company.
Other
- Bachelor's Degree Computer Science, Computer Engineering, Software Engineering, or other related technical field.
- Master's Degree Computer Science, Computer Engineering, Software Engineering, or other related technical field.
- Excellent organizational, prioritization and analytical abilities.
- Have proven experience working in incremental execution through successful launches.
- Experience working in agile environment.