Building foundation models at Apple, including infrastructure, datasets, and models with fundamental general capabilities such as understanding and generation of text, images, speech, videos, and other modalities and apply these models to Apple products.
Requirements
- Proven track record in training or deployment of large models or building large-scale distributed systems.
 
- Proficient programming skills in Python and one of the deep learning toolkits such as JAX, PyTorch, or Tensorflow
 
- Web-scale information retrieval
 
- Human-like conversation agent
 
- Multi-modal perception for existing products and future hardware platforms
 
- On-device intelligence and learning with strong privacy protections
 
Responsibilities
- building infrastructure, datasets, and models with fundamental general capabilities such as understanding and generation of text, images, speech, videos, and other modalities
 
- apply these models to Apple products
 
- tackle some of the most challenging problems in foundation models and deep learning, including natural language processing, multi-modal understanding, and combining learning with knowledge
 
- building systems that push the frontier of deep learning in terms of scaling, efficiency, and flexibility
 
- identify and develop novel applications of deep learning in Apple products
 
- training or deployment of large models
 
- building large-scale distributed systems
 
Other
- Ability to work in a collaborative environment
 
- PhD, or equivalent practical experience, in Computer Science, or related technical field.
 
- passionate about building systems
 
- delight millions of users in Apple products
 
- See your ideas not only published in papers, but also improve the experience of millions of users.