Develop enterprise grade data pipelines for clients, support reporting, business intelligence, and data science end users through ETL and ELT operations, and contribute to the growth of the Data Exploitation Practice.
Requirements
- Python
- SQL
- ETL Tools: Python, Informatica, Pentaho, Talend
- Big data tools: Hadoop, Spark, Kafka, etc.
- Relational SQL and NoSQL databases, including Postgres, CloudSQL, MongoDB
- Data pipeline and workflow management tools: Azkaban, Luigi, Airflow, etc.
- AWS cloud services: EC2, EMR, RDS, Redshift (or Azure and GCP equivalents)
Responsibilities
- Assess and understand ETL jobs and workflows
- Create reusable data pipelines from source to target systems
- Test, validate, and deploy ETL pipelines
- Support reporting, business intelligence, and data science end users through ETL and ELT operations
- Work with data architects to create data models and design schemas for RDBMS, warehouse, and data lake systems
- Key must have skill sets – Python, SQL
- Work within an Agile software development lifecycle
Other
- Ability to hold a position of public trust with the US government.
- 2-4 years industry experience coding commercial software and a passion for solving complex problems
- Experience supporting project teams of developers and data scientists who build web-based interfaces, dashboards, reports, and analytics/machine learning models
- As part of the application process, you are expected to be on camera during interviews and assessments.
- We reserve the right to take your picture to verify your identity and prevent fraud.