The company is looking to solve complex and critical challenges by analyzing and acting on data insights.
Requirements
- Experience with common data science toolkits, such as Python, PySpark, R
- Excellent understanding of machine learning techniques and algorithms
- Knowledge of cloud architectures, including Azure, AWS, or GCP
- Advanced SQL skills, including CTEs and window functions
- Distributed systems knowledge, especially of HDFS and the Hadoop ecosystem
- Familiarity with at least one ETL tool, such as Alteryx, KNIME, SSIS, Pentaho, or DataStage
- Knowledge of any of the following languages: Python, JavaScript, C, or similar
Responsibilities
- Create ETL workflows, scripts, statistical models, and visualizations
- Design, build, test, execute, and support data migration, cleansing, wrangling, etc.
- Select features, build and optimize classifiers using machine learning techniques
- Execute machine learning projects using state-of-the-art methods
- Create automated anomaly detection systems and track performance
- Build effective, reliable, and robust ETL processes
- Design database models, consistent table structures, and advanced dimensional schemas
Other
- Bachelor’s degree with concentration in Computer Science, Engineering or another quantitative field
- 7+ years of applicable professional experience
- Data-oriented personality
- Ability to synthesize requests from team members at client sites
- Desire to actively engage in geographically dispersed teams
- Capability to be a creative, innovative problem solver
- Strong verbal and written communication skills in English
- Ability and willingness to work long hours and travel if necessary
- Ability to work full-time in an office and remote environment