Speechify is looking to improve the performance, latency, throughput, and efficiency of its deployed AI models.
Requirements
- Experience shipping Python-based services
- Experience with public cloud environments, GCP preferred
- Experience with Infrastructure such as Code, Docker, and containerized deployments
- Preferred: Experience deploying high-availability applications on Kubernetes
- Preferred: Experience deploying ML models to production
Responsibilities
- Work alongside machine learning researchers, engineers, and product managers to bring AI Voices to customers for a diverse range of use cases
- Deploy and operate the core ML inference workloads for AI Voices serving pipeline
- Introduce new techniques, tools, and architecture that improve the performance, latency, throughput, and efficiency of deployed models
- Build tools to give visibility into bottlenecks and sources of instability and then design and implement solutions to address the highest priority issues
Other
- Work ethic, solid communication skills, and obsession with winning are paramount
- Autonomy, fostering focus and creativity
- The opportunity to have a significant impact in a revolutionary industry
- Competitive compensation, a welcoming atmosphere, and a commitment to an exceptional asynchronous work culture