Developing new techniques for understanding and steering large AI models to unlock the next frontier of safe and powerful foundation models.
Requirements
- Deep familiarity with large models and a passion for understanding how they work
- Fluency in Python and ML frameworks such as PyTorch
- Experience leading research or contributing to open-source codebases
- Familiarity with interpretability, alignment, or safe model development
- Experience in startup or fast-paced lab environments
Responsibilities
- Conduct original research in interpretability and related fields
- Prototype techniques to visualize and manipulate internal model structures
- Collaborate with engineering to turn research into production-ready tools
- Share work through publications, demos, and open-source contributions
- Help define and evolve research direction
Other
- Put mission and team first
- Improve constantly
- Take ownership and initiative
- Action today
- PhD or equivalent experience in ML, computer science, or a quantitative science
- Strong writing and communication skills for explaining complex ideas
- Drive to move quickly and take ownership