Susquehanna is looking to build reliable, high-performance systems for handling large and complex datasets to power their AI research
Requirements
- Experience building and maintaining data pipelines and ETL systems at scale
- Experience with large-scale ML infrastructure and familiarity with training and inference workflows
- Strong understanding of best practices in data management and processing
- Knowledge of systems level programming and performance optimization
- Proficiency in software engineering in python
- Understanding of AI/ML workloads, including data preprocessing, feature engineering, and model evaluation
Responsibilities
- Design and implement high-performance data pipelines for processing large-scale datasets with an emphasis on reliability and reproducibility
- Collaborate with researchers to translate their requirements into scalable, production-grade systems for AI experimentation
- Optimize resource utilization across our distributed computing infrastructure through profiling, benchmarking, and systems-level improvements
- Implement low-latency high-throughput sampling for models
Other
- Partner with researchers to support experimental workflows
- Help translate evolving needs into efficient, production-ready solutions
- Collaborative culture
- Innovation is at the core of everything we do