Gridware is looking to build and maintain the pipelines and data systems powering their Active Grid Response platform to monitor the electrical, physical, and environmental aspects of the grid that affect reliability and safety.
Requirements
- Strong proficiency in Python and SQL
- Familiarity with data warehouses, Lakehouse platforms, or big data tools (Databricks, Spark, or equivalent)
- Experience with pipeline orchestration tools (Airflow, Dagster, Prefect, etc.)
- Understanding of event-driven systems or streaming platforms (Kafka, Kinesis, Pub/Sub)
- Solid foundation in data modeling, testing, and version control
- Experience with IoT, telemetry ingestion, or time-series data
- Understanding of Protobuf, Avro, Parquet, or serialization formats
Responsibilities
- Building ETL/ELT pipelines that ingest transformer, pole, and sensor telemetry into Gridware’s Data Lake and Lakehouse
- Developing and maintaining real-time and batch ingestion processes using Python, SQL, Databricks, and Spark
- Implementing data quality checks, validation rules, and automated testing for stable operations
- Collaborating with Software, Firmware, and Data Science teams to define ingestion schemas and transformations
- Working with cloud-native tools to optimize pipeline throughput and cost efficiency
- Monitoring pipelines for reliability, troubleshooting issues, and contributing to on-call rotations
- Writing documentation for data processes, models, and metadata
Other
- 2–4 years of experience as a Data Engineer (or Backend Engineer with heavy data exposure)
- Ability to work collaboratively in a high-autonomy, fast-paced environment
- Health, Dental & Vision (Gold and Platinum with some providers plans fully covered)
- Paid parental leave
- Alternating day off (every other Monday)