Red Hat Inference engineering team accelerates AI for the enterprise and brings operational simplicity to GenAI deployments. As leading maintainers of the vLLM and LLM-D projects, and inventors of state-of-the-art techniques for model quantization and sparsification, our team provides a stable platform for enterprises to build, optimize, and scale LLM deployments.
Requirements
- 5+ years of significant hands-on software development and system design experience.
- 2+ years of managing software engineering teams.
- Proven experience in leading machine learning engineering teams, with a track record of successful project delivery and development of software engineers.
- Experience in machine learning frameworks and tools, such as PyTorch, and HuggingFace.
- Excellent programming skills in languages like Python, or C++/CUDA.
- Experience with developing and scaling applications with Kubernetes.
- Solid understanding of core machine learning concepts.
Responsibilities
- Lead and inspire a distributed team of individual contributors, fostering a collaborative and innovative work environment.
- Engage with the AI and machine learning open source communities such as vLLM, llm-d, and several other open source communities.
- Work with product management and engineering teams to develop technology roadmaps and schedules, and communicate these schedules externally.
- Work with cross-functional engineering managers and teams on documentation, product management, and quality assurance to coordinate tasks necessary for releasing enterprise-quality MLOps software.
- Working closely with the technical leads and scrum leads to direct the team in agile development.
- Mentor and nurture team members in their career development, and professional growth.
- Recruit and build a world class engineering team.
Other
- Excellent written and verbal communication skills.
- Ability to lead and work with diverse and distributed teams from multiple countries and cultures.