SMX is seeking a Data Engineer III to design, implement, and maintain scalable, secure, and performant enterprise-scale data infrastructure and analytics systems for government and commercial clients.
Requirements
- Advanced proficiency in SQL and strong programming skills in Python, Scala, or Java
- Expertise in Git, collaborative development, and shell scripting
- Hands-on experience with big data processing (Apache Spark, Hadoop) and workflow orchestration (Airflow, Step Functions)
- Experience with streaming technologies (Kafka, Kinesis) and data serialization formats (Parquet, Avro)
- Practical experience with AWS, Azure, or Google Cloud platforms and their data services
- Proficient with containerization (Docker) and orchestration (Kubernetes)
- Knowledge of IaC (Terraform) and cloud security best practices
Responsibilities
- Design and build automated data pipelines (ingestion, transformation, standardization) from diverse sources, including streaming platforms
- Optimize pipelines for performance and reliability; develop reusable components and frameworks
- Ensure robust data lineage, metadata management, monitoring, and logging across all systems
- Manage cloud data infrastructure (compute, storage, networking) utilizing Infrastructure-as-Code (IaC) principles
- Design, implement, and optimize database schemas for all workloads (transactional, analytical, NoSQL)
- Establish and maintain high availability, backup, and disaster recovery strategies
- Implement automated testing (unit, integration, data quality) and validation rules
Other
- Ability to obtain Secret security clearance (US citizenship required)
- 5-7 years of hands-on data engineering experience; 2-3 years in a senior or technical lead role
- Proven track record designing and implementing production-grade, enterprise-scale data systems
- Experience in government contracting or with sensitive data (preferred)
- Understanding of data security, encryption, and compliance frameworks (NIST, FedRAMP)