BNSF Railway is looking to design, develop, and maintain scalable data streaming solutions and data pipelines to support real-time and batch processing, enabling data-driven decision-making and operational efficiency.
Requirements
- Proficiency in programming languages such as Python, SQL, and Java
- Strong understanding of data warehousing concepts, including dimensional modeling and star schema
- Experience with ETL processes and cloud-based solutions
- Familiarity with database management systems (SQL, NoSQL) and lakehouse technologies (Apache Iceberg, Delta Lake, Apache Trino)
- Experience with data streaming technologies like Apache Kafka, AWS Kinesis, Apache Flink, and Apache Spark
- Minimum of 2 years of experience in software engineering, data engineering, or related fields
Responsibilities
- Design and implement efficient data streaming solutions to process real-time data
- Develop and maintain robust data pipelines for real-time and batch processing, including extraction, transformation, and loading (ETL)
- Create and manage large-scale data warehouses to support business intelligence and analytics
- Integrate data from diverse sources to produce certified datasets ensuring accuracy and reliability
- Optimize data workflows, monitor pipelines, and troubleshoot issues to ensure high performance and reliability
- Ensure data security and compliance with industry standards and regulations
- Write efficient, well-documented code with proper unit tests and documentation
Other
- Authorized to work in the US
- Excellent problem-solving skills and attention to detail
- Strong communication and collaboration skills
- Ability to learn new technical concepts quickly and adapt to changing environments
- Collaborate with data scientists, analysts, and stakeholders to gather data requirements and deliver solutions