Walmart is looking to leverage advanced AI and data engineering to build scalable, intelligent data solutions for advanced analytics, forecasting, and operational reporting in a multi-cloud retail environment.
Requirements
- Strong proficiency in Python, PySpark, SQL, and shell scripting.
- Hands-on experience with GCP services: Dataproc, BigQuery, Cloud Functions, Cloud Dataflow, Cloud Storage, Vertex AI.
- Familiarity with Azure Data Factory, Snowflake, and multi-cloud orchestration.
- Experience with CI/CD tools: Jenkins, GitHub, Azure DevOps.
- Solid understanding of data warehousing, ETL processes, and data modeling.
- Practical experience with LangChain, LangGraph, RAG architectures, and Agentic AI systems.
- Google Professional Data Engineer certification is a plus.
Responsibilities
- Design, build, and optimize data pipelines using PySpark, Apache Spark, and Airflow for orchestration.
- Develop and deploy workflows on Google Cloud Platform (GCP) using Dataproc, Cloud Scheduler, Cloud Storage, and BigQuery.
- Implement multi-cloud data architectures integrating Azure, GCP, and Snowflake for data lakes and warehouses.
- Ensure data quality, lineage, and governance using tools like Apache Hudi, GitHub Actions, and Terraform.
- Build and deploy RAG-based AI solutions for intelligent document retrieval and contextual response generation.
- Integrate LangChain and LangGraph frameworks to orchestrate complex AI workflows and multi-agent systems.
- Develop Agentic AI capabilities to automate decision-making and enhance data-driven insights.
Other
- 5+ years of experience in data engineering/software engineering, with at least 4 years on GCP.
- Demonstrated ability to work with integrity, respect for individuals, and a customer-first mindset.
- Proven track record of collaboration, continuous learning, and striving for excellence.
- Bachelor’s or Master’s degree in Computer Science, Engineering, or related field.
- Knowledge of retail domain data flows and key performance indicators (KPIs).