CompoSecure is looking to leverage data to support business analytics and reporting, requiring robust data pipelines and solutions for enterprise analytics platforms like Power BI.
Requirements
- Strong expertise in data languages (Python, SQL, DAX, M, R, etc.) and handling large data sets using PySpark.
- Strong expertise in Azure Data Factory, Azure Synapse Analytics, Databricks (nice-to-have), and Microsoft Fabric.
- Hands-on experience implementing Medallion architecture in a lakehouse environment.
- Expertise in transforming Parquet datasets into Delta tables for reliable, incremental processing.
- Experience building ETL/ELT pipelines and working with both structured and unstructured data.
- Experience with Power BI integration and data modeling.
- Familiarity with CI/CD for data pipelines a plus.
Responsibilities
- Design, develop, and maintain data pipelines using Azure Data Factory, Azure Synapse, and Microsoft Fabric Dataflows.
- Implement and maintain Medallion architecture in Fabric and Azure environments.
- Build ingestion and transformation flows that efficiently convert raw Parquet files into Delta tables to support curated, incremental, and governed datasets optimized for Power bi.
- Implement efficient data lakehouse patterns within Microsoft Fabric.
- Design and maintain star schemas, dimensional models, and semantic layers for Power Bi.
- Tune queries, optimize partitioning, and manage performance across Fabric and Synapse environments.
- Implement data governance, lineage, and cataloging via Microsoft Purview and Fabric capabilities.
Other
- Minimum 5 years of relevant experience.
- Excellent analytical and problem-solving skills.
- Excellent record keeping and auditing skills.
- Excellent communication skills, both oral and written.
- Must be able to work beyond 40 hour week during upgrades or new functionality rollout if necessary.