The company is looking to solve the problem of migrating and optimizing SAS programs and workflows to PySpark, ensuring data security, privacy, and compliance with organizational standards.
Requirements
- Experience in SAS and must have experience in Base SAS and SAS EG (Enterprise Guild) with exposure to Python and PySpark is good to have.
- Experience in SAS Mainframe (Good to have).
- Strong understanding of data architecture, data modelling, and ETL/ELT processes.
- Experience with big data processing and distributed computing systems like Spark and Kafka.
- Proficient in Python and Scala scripting for data manipulation and analysis.
- Proficiency in SAS programming and tools
- Proven expertise in Spark, SQL, including data pipeline and notebook development.
Responsibilities
- Experience implementing data integration workflows using SAS tools such as SAS DI Studio.
- Sanitize SAS programs in preparation for conversion.
- Optimize SAS programs in preparation for conversion.
- Incorporate changes in accelerator to ensure generated code aligns to required standards
- Utilize accelerator to convert SAS programs to PySpark programs.
- Demonstrate and document code lineage.
- Unit Test generated PySpark programs
Other
- 6+ years of experience
- Excellent communication, collaboration, and problem-solving skills.
- Strong leadership and team-building abilities.
- Ensure compliance with data security, privacy regulations, and organizational standards.
- Prepare for deployment.