Developing the core PyTorch 2.0 technologies, innovating and advancing the state-of-the-art of ML compilers, and accelerating PT2 adoption through direct engagements with OSS and industry users at Meta
Requirements
- Research or industry experience in developing compilers, ML systems, ML accelerators, GPU performance, and similar
- Advanced in Python or C++ programming
- Experience in developing PyTorch/PT2, Triton, MLIR, JAX, XLA, TVM is a huge plus
- Knowledge in GPU architecture, ML accelerator performance, and developing high-performance kernels
- Experience in building OSS communities and extensive social media presence in the ML Sys domain
- Experience with training models, end-to-end model optimizations, or applying ML to systems
- Knowledge of communication collectives, PyTorch distributed, and parallelism
Responsibilities
- Develop the PT2 compiler (e.g., TorchDynamo, TorchInductor, PyTorch Distributed, PyTorch Core)
- Improve PyTorch performance via systematic solutions for the entire community
- Explore the intersection of the PyTorch compiler and PyTorch distributed
- Optimize Generative AI models across the stack (pre-training, fine-tuning, and inference)
- Collaborate with users of PyTorch to enable new use cases of PT2 technologies both inside and outside Meta
Other
- Currently has, or is in the process of obtaining a Bachelor's degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience
- Currently has or is in the process of obtaining a PhD degree in Computer Science, Computer Engineering, relevant technical field, or equivalent practical experience
- Individual compensation is determined by skills, qualifications, experience, and location
- Meta offers benefits
- Meta is proud to be an Equal Employment Opportunity employer