Speechify aims to solve the problem of reading being a barrier to learning by providing text-to-speech products that allow users to consume content through audio. The company is looking for someone to deploy and operate ML inference workloads for their AI Voices serving pipeline, improving performance, latency, throughput, and efficiency of deployed models.
Requirements
- Experience shipping Python-based services
- Experience being responsible for the successful operation of a critical production service
- Experience with public cloud environments, GCP preferred
- Experience with Infrastructure such as Code, Docker, and containerized deployments.
- Preferred: Experience deploying high-availability applications on Kubernetes.
- Preferred: Experience deploying ML models to production
Responsibilities
- Work alongside machine learning researchers, engineers, and product managers to bring our AI Voices to their customers for a diverse range of use cases
- Deploy and operate the core ML inference workloads for our AI Voices serving pipeline
- Introduce new techniques, tools, and architecture that improve the performance, latency, throughput, and efficiency of our deployed models
- Build tools to give us visibility into our bottlenecks and sources of instability and then design and implement solutions to address the highest priority issues
Other
- This is a key role and ideal for someone who thinks strategically, enjoys fast-paced environments, passionate about making product decisions, and has experience building great user experiences that delight users.
- Work ethic, solid communication skills, and obsession with winning are paramount.
- Competitive compensation, a welcoming atmosphere, and a commitment to an exceptional asynchronous work culture
- The privilege of working on a product that changes lives, particularly for those with learning differences like dyslexia, ADD, and more
- An active role at the intersection of artificial intelligence and audio – a rapidly evolving tech domain