Toyota’s Data Department is looking to build state of the art Data Platforms and develop Data Products in support of various Analytical needs across the organization.
Requirements
- Hands-on experience with cloud platforms such as AWS, Microsoft Azure, or Google Cloud Platform (GCP)
- 1-2 years of developing Infrastructure as Code (IaC) tools with Terraform, AWS CloudFormation, or Azure ARM template for provisioning and managing cloud resources relevant to data platforms (e.g., S3, EC2, Lambda, RDS, SNS, Restful APIs, CloudWatch etc)
- 1-2 yrs of experience developing data pipelines using Spark, Scala, Python, Databricks or Spark platforms
- Experience writing and optimizing SQL queries for data analysis, extraction, transformation, and loading
- Understanding of cloud security fundamentals, including IAM roles, policies, and encryption algorithms
- Exposure to monitoring & alerting solutions for platform availability, data pipelines, infrastructure health checks, performance, and cost observability
- Knowledge of Python, Bash, or PowerShell and building DevOps solutions with GitHub, Jenkins, Ansible, or similar
Responsibilities
- Develop, and maintain scalable, secure, and cost-optimized cloud infrastructure for Data Lake, Data Warehouse, and analytics platforms using Infrastructure as Code (IaC) tools such as Terraform, AWS CloudFormation, or equivalent.
- Develop automations for provisioning, configuration, and deployment of cloud resources (e.g., AWS S3, Redshift, Databricks, Sagemaker, EMR, Lambda, EC2, SNS, Elastic Cache) to support data engineering workloads.
- Build and maintain CI/CD pipelines for data platform components and data engineering workflows using Jenkins, Ansible, Chef, XL Release, XL Deploy, or similar tools.
- Monitor system performance, troubleshoot incidents, and implement continuous improvements to enhance platform reliability and scalability.
- Build and maintain platform health checks, monitoring, alerting, and resiliency mechanisms to ensure high availability and performance of data systems.
- Develop, and maintain end-to-end data engineering pipelines using Apache Spark, Scala, Databricks, and Informatica ETL tools to ingest, transform, and load large-scale datasets.
- Develop automation frameworks for pipeline testing, validation & Observability solutions to track data quality, timeliness, and accuracy.
Other
- Bachelor’s degree in Information Systems, Computer Science, or related discipline or equivalent work experience
- Strong problem-solving skills, adept with documenting processes, workflows, and technical solutions clearly
- Good communication skills to collaborate effectively with cross-functional teams including product owners, architects, and developers
- Toyota does not offer sponsorship of job applicants for employment-based visas or any other work authorization for this position at this time.
- Applicants for our positions are considered without regard to race, ethnicity, national origin, sex, sexual orientation, gender identity or expression, age, disability, religion, military or veteran status, or any other characteristics protected by law.