WillHire is looking to solve the problem of discovering exceptional talent in the Data Engineering vertical by expanding its Big Data Engineering Internship Cohort.
Requirements
- Strong understanding of databases, SQL, and data modeling concepts
- Familiarity with Python/Java/Scala for data processing
- Basic knowledge of data warehousing, ETL concepts, and data pipelines
- Understanding of cloud platforms (AWS/GCP/Azure) and their data services
- Certifications such as AWS Certified Data Engineer, Google Cloud Data Engineer, or Microsoft Azure Data Fundamentals (even if in-progress)
- Experience with big data tools like Hadoop, Spark, or Kafka
- Familiarity with data visualization tools (Tableau, Power BI, Looker)
Responsibilities
- Assist in designing, developing, and maintaining ETL pipelines for structured and unstructured data.
- Work with databases (SQL/NoSQL) to ensure data accuracy, integrity, and accessibility.
- Support integration of APIs and third-party data sources into the data ecosystem.
- Help optimize data storage, transformation, and retrieval for performance and scalability.
- Contribute to data quality checks, validation processes, and monitoring frameworks.
- Assist in developing dashboards and reports for business and product stakeholders.
- Research new tools, frameworks, and best practices in data engineering & analytics.
Other
- Currently pursuing (or recently completed) B.Tech/BE/M.Tech/MSc in Computer Science, Data Engineering, IT, or related fields
- Critical thinking, attention to detail, and a proactive learning mindset
- Motivation, culture fit, and communication skills
- Resume + short write-up on your data engineering interests/projects
- Participation in data hackathons, Kaggle competitions, or open-source data projects