At Google DeepMind, we’re working to advance the state of the art in artificial intelligence, ensuring safety and ethics are the highest priority, particularly in scenarios where the agent’s capabilities may exceed those of the systems tasked with monitoring it.
Requirements
- Ph.D. in Computer Science or related quantitative field, or B.S./M.S. in Computer Science or related quantitative field with 5+ years of relevant experience.
 
- Demonstrated research or product expertise in AI safety, AI alignment, or a related security field.
 
- Experience with adversarial research, red-teaming, or vulnerability research, particularly in complex software or AI/ML systems.
 
- Strong software engineering skills and experience with ML frameworks like JAX, PyTorch, or TensorFlow.
 
- Familiarity with concepts from game theory or mechanism design problems as they apply to AI.
 
- A track record of landing research impact within multi-team collaborative environments.
 
Responsibilities
- Go beyond traditional security assumptions to model how a highly capable agent could misuse its access, exfiltrate data, or establish a rogue deployment within complex production environments.
 
- Develop techniques for monitoring advanced agents. How can we detect emergent deception, collusion, or obscured long-term plans before they result in harmful actions?
 
- Create novel evaluation methodologies and benchmarks to measure the effectiveness of different control strategies against highly capable simulated adversaries.
 
- Identify and formalize unsolved research problems in AI control, focusing on the unique challenges posed by agents that may exceed human oversight capabilities.
 
- Design, prototype, and evaluate novel control systems and monitoring techniques. This includes theoretical work, large-scale experiments, and building proof-of-concept systems.
 
- Collaborate closely with teams working on Gemini and agent infrastructure to understand emergent risks and integrate control mechanisms directly into the systems where they are most needed.
 
- Publish groundbreaking research and contribute to the broader academic and policy conversation on long-term AI safety and control.
 
Other
- Ph.D. in Computer Science or related quantitative field, or B.S./M.S. in Computer Science or related quantitative field with 5+ years of relevant experience.
 
- We value diversity of experience, knowledge, backgrounds and perspectives and harness these qualities to create extraordinary impact.
 
- We are committed to equal employment opportunity regardless of sex, race, religion or belief, ethnic or national origin, disability, age, citizenship, marital, domestic or civil partnership status, sexual orientation, gender identity, pregnancy, or related condition (including breastfeeding) or any other basis as protected by applicable law.
 
- If you have a disability or additional need that requires accommodation, please do not hesitate to let us know.
 
- A creative and rigorous researcher who is passionate about tackling the foundational safety challenges of advanced AI.