Corning is looking to improve production performance by designing, running, and scaling pipelines across cloud, on-prem, and edge environments to deliver reliable, actionable insights
Requirements
- Strong Python and SQL; familiarity with Java, C, or JavaScript
- Hands-on with ETL frameworks (Databricks/Spark SQL, AWS Glue, Informatica)
- Experience with AWS/Azure/GCP for data storage and compute
- Exposure to MES/SCADA and industrial data protocols
- CI/CD tools (GitLab, Jenkins, Azure DevOps) and large-scale data/analytics systems
- Time-series, sensor, or IoT data experience
- Data visualization tools (Power BI, Tableau)
Responsibilities
- Design and build robust data pipelines to collect real-time data from production lines and IoT devices
- Build and optimize ETL/ELT pipelines for MES, SCADA, PLCs, metrology/test systems, and IoT sensors
- Automate data workflows, monitoring, and CI/CD to improve resiliency and reduce manual effort
- Manage and monitor data environments with quality checks, regression testing, and performance tuning for pipelines and data lakes
- Collaborate with domain experts, controls engineers, and data teams to define requirements and deliver solutions
- Define IT/OT data interfaces; ingest, maintain, and enhance industrial data sources
- Provision and automate infrastructure using Terraform, Ansible, Kubernetes; containerize with Docker
Other
- BS/MS in Computer Science, Engineering, Information Systems, or related field
- 3+ years in data engineering (manufacturing/industrial automation preferred)
- Strong problem-solving, clear communication, and collaboration in a matrixed environment
- Proactive, adaptable, committed to documentation and knowledge sharing
- This position does not support immigration sponsorship