Altimate AI is tackling the overwhelming volume of data and shortage of skilled data professionals in modern enterprise analytics by automating and accelerating data operations.
Requirements
- Expert proficiency in Python
- Deep, hands-on experience with the modern data stack, including: Cloud platforms (AWS, GCP, or Azure)
- Data Warehouses/Lakes (Snowflake, Databricks, or similar)
- Data Transformation/Modeling (DBT)
- Workflow Orchestration (Airflow or similar)
Responsibilities
- Lead the technical implementation, integration, and optimization of Altimate AI's platform within complex enterprise data environments.
- Serve as the primary technical point of contact for key accounts, translating customer needs into technical requirements for the product and engineering teams.
- Develop custom scripts, integrations, and tools using Python to extend the capabilities of our platform and ensure seamless operation within customer ecosystems.
- Troubleshoot, diagnose, and resolve complex technical issues related to data connectivity, performance, and workflow integration.
- Contribute to the core product by identifying common integration patterns and developing reusable components or product features.
- Conduct technical workshops and provide expert guidance on data best practices, leveraging tools like Snowflake, Databricks, and dbt.
Other
- 4+ years of professional software engineering or data engineering experience
- Bachelor's or Master's degree in Computer Science or a related technical field
- Experience in a Forward Deployed, Solutions Architect, or Customer-Focused Engineering role for a B2B SaaS product.
- Familiarity with Data Governance, Data Quality, and Data Observability principles.
- Strong communication skills, with the ability to articulate complex technical concepts to both technical and non-technical audiences.