Job Board
LogoLogo

Get Jobs Tailored to Your Resume

Filtr uses AI to scan 1000+ jobs and finds postings that perfectly matches your resume

D&H Distributing Logo

Machine Learning Engineer

D&H Distributing

Salary not specified
Oct 16, 2025
Harrisburg, PA, US
Apply Now

D&H is looking to operationalize data science and AI models to drive insights, automation, and customer value by building end-to-end machine learning pipelines that scale across their new cloud-based data architecture.

Requirements

  • Proficiency in Python, SQL, PySpark and ML libraries (scikit-learn, TensorFlow, PyTorch).
  • Hands-on experience with Databricks MLflow and/or Azure ML pipelines.
  • Strong understanding of data modeling, feature engineering, and ML lifecycle management.
  • Knowledge of cloud architectures (Azure, Databricks, Snowflake) and modern data engineering practices.
  • Experience integrating ML with SAP Datasphere/SAC and embedding outputs into BI dashboards.
  • Familiarity with forecasting, anomaly detection, NLP, or recommender systems.
  • Knowledge of DevOps/MLOps tooling (CI/CD pipelines, Docker, Kubernetes).

Responsibilities

  • Build, deploy, and monitor ML models using Databricks MLflow, and Python frameworks (scikit-learn, TensorFlow, PyTorch).
  • Automate model training, retraining, and deployment pipelines integrated with Databricks Delta Lake and Azure Data Lake.
  • Implement CI/CD for ML models to ensure reliable and scalable deployment.
  • Develop feature engineering pipelines that leverage Azure Synapse, Databricks, and SAP Datasphere.
  • Optimize data retrieval and transformation for real-time or batch ML use cases.
  • Establish model monitoring and drift detection to maintain accuracy and business relevance.
  • Embed ML outputs into Tableau, Power BI, SAP Analytics Cloud (SAC), or API-driven integrations.

Other

  • This is a fully remote role.
  • 3+ years in machine learning engineering or applied ML roles.
  • Collaborate with data engineers, analysts, and business stakeholders to translate requirements into ML solutions.
  • Support citizen data science initiatives by building reusable frameworks and tools.
  • Exposure to Netezza and Cognos legacy environments (for migration and feature replication).