IMC is seeking a Data Engineer to architect, develop, and deploy their Big Data environment and data processing pipelines.
Requirements
- 3+ years of experience building Kafka streaming applications and/or maintaining Kafka clusters
- 2+ years of experience building applications/pipelines with Big Data backends (S3, HDFS, Databricks, Iceberg, etc)
- Experience with Apache Spark, Apache Flink or similar tools
- Strong Java, Python, and SQL development skills
- Experience with common data-science toolkits, especially python-based
- Hands-on experience with Kubernetes and Docker
- Experience with monitoring tools such as Prometheus/Grafana, Alert Manager, Alerta and OpsGenie
Responsibilities
- Architect, develop and deploy our Big Data environment (Kafka, Hadoop, Dremio, etc.)
- Build, deploy, and monitor our data processing pipelines (Java, Python, Spark, Flink)
- Collaborate with development teams on data modeling, data ingestion, and capacity planning
- Work with users to ensure data integrity and availability
- Act as a Big Data SME and consult on a variety of data-related questions from users and developers
Other
- 5+ years experience working in a mature data engineering environment
- Demonstrated ability to troubleshoot and conduct root-cause analysis
- Unix scripting experience (bash, python, etc.)