Apple is looking for researchers to build foundation models with general capabilities in text, images, speech, and other modalities, and apply these models to Apple products to improve user experience.
Requirements
- Demonstrated expertise in deep learning with a publication record in relevant conferences (e.g., NeurIPS, ICML, ICLR, CVPR, ICCV, ECCV, KDD, ACL, ICASSP, InterSpeech) or a track record in applying deep learning techniques to products
- Proficient programming skills in Python and one of the deep learning toolkits such as JAX, PyTorch, or Tensorflow
- Web-scale information retrieval
- Human-like conversation agent
- Multi-modal perception for existing products and future hardware platforms
- On-device intelligence and learning with strong privacy protections
Responsibilities
- building foundation models
- build infrastructure, datasets, and models with fundamental general capabilities such as understanding and generation of text, images, speech, videos, and other modalities
- apply these models to Apple products
- developing algorithms, techniques, and systems that push the frontier of deep learning
- tackle some of the most challenging problems in foundation models and deep learning, including natural language processing, multi-modal understanding, and combining learning with knowledge
- identify and develop novel applications of deep learning in Apple products
Other
- PhD, or equivalent practical experience, in Computer Science, or related technical field.
- Ability to work in a collaborative environment