At Schwab, the business problem is to transform people's lives by changing the way the world invests, and the technical problem is to re-architect and rebuild the core enterprise data platform to support the rapidly growing and evolving business.
Requirements
- Expertise in ETL development, Data Quality, Best Practices, Exception Handling
- SQL, Google Cloud Products including BigQuery, Cloud Storage, Dataflow, Composer, Pub/Sub, SQL Server Integration Services
- Expertise in Python coding skills
- Good understanding of Architecture and Development Standards and Guidelines
- Experience working with and sourcing data from Corporate Systems like Workday, BMC Remedy Helix, Jira, Rally, etc.
- Strong understanding of Database and SQL
- Experience in dev tools such as GitHub, Bamboo, Liquibase
Responsibilities
- Have expertise in data pipeline development.
- Experience working in Agile development environment supporting organization priorities and innovation.
- Work closely with technical lead, development team members to decompose stories, design features, and prioritize tasks.
- Create and maintain artifacts for production support and deployment processes and ensure smooth hand off to support teams.
- Work with onshore and offshore developers and application support to review and improve processes.
- Design, implement, and optimize end-to-end data pipelines for ingesting, processing, and transforming large volumes of structured and unstructured data.
- Develop robust ETL (Extract, Transform, Load) processes to integrate data from diverse sources into our data ecosystem.
Other
- Master's or bachelor's degree in computer science, Information Technology or similar/equivalent area of study
- Minimum of 3-5 years of solid data engineering experience.
- Good verbal and written communication skills required for composing and delivering technical documentation to various levels of technical management.
- Willingness to learn new skills and keep up with technology changes in future to meet business needs.
- Collaborate with analysts, and other stakeholders to understand data requirements and deliver tailored solutions.