Comcast is looking to develop data structures and pipelines to organize, collect, standardize, and transform data to generate insights and address reporting needs.
Requirements
- AWS (including S3, Redshift)
- PySpark
- Databricks
- Big Data Architecture
- Python, SQL
- Apache Spark
- Data Modeling & Pipeline Design
Responsibilities
- Develops data structures and pipelines aligned to established standards and guidelines.
- Ensures data quality during ingest, processing, and final load to target tables.
- Creates standard ingestion frameworks for structured and unstructured data.
- Creates standard methods for end users and downstream applications to consume data, including: Database views, Extracts, Application Programming Interfaces (APIs).
- Develops and maintains information systems (e.g., data warehouses, data lakes), including data access APIs.
- Implements solutions via data architecture, data engineering, or data manipulation on: On-prem platforms (e.g., Kubernetes, Teradata), Cloud platforms (e.g., Databricks).
- Determines appropriate storage platforms across on-prem (minIO, Teradata) and cloud (AWS S3, Redshift) based on privacy, access, and sensitivity requirements.
Other
- 5-7 Years Relevant Work Experience
- Bachelor's Degree (or combination of coursework and experience, or extensive related professional experience)
- Comcast is an equal opportunity workplace.