Hedra is seeking a Research Scientist to push the limits of long-form video generation, focusing on auto-regressive modeling, causal attention mechanisms, and efficient sequence handling to build a multimodal creation platform capable of control, emotion, and creative intelligence.
Requirements
- Deep understanding of transformer architectures, attention mechanisms, and auto-regressive modeling.
- Experience with long-context processing and memory-efficient computation.
- Proficiency in Python and PyTorch; ability to rapidly prototype and iterate on new architectures.
- A record of impactful research or large-scale system deployments.
- Deep understanding of temporal modeling in generative AI and experience building scalable architectures for multi-minute coherent video outputs.
Responsibilities
- Design and implement long video generation architectures, with emphasis on auto-regressive generation, causal attention, and memory-efficient transformer designs.
- Develop methods for maintaining temporal and semantic coherence over long time horizons.
- Work closely with engineering to integrate research into production-grade pipelines.
- Stay on top of recent advances in long-context transformers, sequence compression, and scalable video generation.
- Present results internally and externally, including possible top-tier conference submissions.
Other
- PhD or strong research/industry experience in Computer Science, Machine Learning, or related fields, with a focus on sequence modeling or generative models.
- We value startup energy, initiative, and the ability to turn bold ideas into real products.
- Our team is fully in-person in SF/NY with a shared love for whiteboard problem-solving.
- We encourage you to apply even if you don't meet every requirement — we value curiosity, creativity, and the drive to solve hard problems.