Hone is seeking a Data Engineering Intern to support the design, development, and maintenance of data systems and pipelines, ensuring data accuracy, reliability, and accessibility for analytics, reporting, and product use cases.
Requirements
- Strong foundational knowledge of SQL and experience querying relational databases.
- Proficiency in Python and a strong interest in distributed data processing (PySpark experience is a plus).
- Understanding of data modeling, data warehousing, or analytics engineering concepts.
- Exposure to or coursework involving data pipeline orchestration or ETL development.
Responsibilities
- Design, build, and maintain scalable data pipelines and ETL processes to support analytics, reporting, and product use cases.
- Integrate data from multiple internal and external sources, ensuring quality, consistency, and reliability.
- Develop and maintain data models, schemas, and warehouses to support BI, ML, and operational needs.
- Collaborate with engineers, analysts, and product teams to translate business requirements into technical data solutions.
- Participate in data quality checks, testing, validation, and performance optimization.
- Monitor, optimize, and troubleshoot data infrastructure for performance and scalability.
- Contribute to data documentation and ensure best practices around data governance, reliability, and scalability.
Other
- Currently pursuing a Bachelor’s or Master’s degree in Computer Science, Data Engineering, Data Science, Information Systems, or a related field.
- Strong analytical thinking, problem-solving abilities, and attention to detail.
- Eagerness to learn new technologies and frameworks, with a focus on self-improvement.
- Effective communication skills and the ability to work collaboratively in a remote cross-functional environment.
- A stable internet connection and access to a PC/laptop.