Meta is seeking Research Scientists to join its Fundamental AI Research (FAIR) organization, focused on making significant advances in AI, specifically in building foundation models for audio understanding and generation, and pushing the frontier of multimodality research.
Requirements
- Research publications reflecting experience in related research fields: audio (speech, sound, or music) generation, text-to-speech (TTS) synthesis, text-to-music generation, text-to-sound generation, speech recognition, speech / audio representation learning, vision perception, image / video generation, video-to-audio generation, audio-visual learning, audio language models, lip sync, lip movement generation / correction, lip reading, etc
- Familiarity with one or more deep learning frameworks (e.g. pytorch, tensorflow, …)
- Experienced in Python programming language
- Publication record at peer-reviewed AI conferences (e.g. ACL, EMNLP, NeurIPS, ICLR, ICML or similar)
Responsibilities
- Develop algorithms based on state-of-the-art machine learning and neural network methodologies
- Perform research to advance the science and technology of intelligent machines
- Conduct research that enables learning the semantics of data across multiple modalities (audio, speech, images, video, text, and other modalities)
- Design and implement models and algorithms
- Work with large datasets, train / tune / scale the models, create benchmarks to evaluate the performance, open source and publish
Other
- 2+ years of experience holding an industry, faculty, academic, or government researcher position
- Must obtain work authorization in the country of employment at the time of hire, and maintain ongoing work authorization during employment