Job Search and Career Advice Platform

Attiva gli avvisi di lavoro via e-mail!

Machine Learning Operations Engineer

United Language Group

Borghetto Lodigiano

In loco

EUR 45.000 - 70.000

Tempo pieno

Oggi
Candidati tra i primi

Genera un CV personalizzato in pochi minuti

Ottieni un colloquio e una retribuzione più elevata. Scopri di più

Descrizione del lavoro

A leading language services company in Italy is seeking a Machine Learning Operations Engineer. The role involves designing and maintaining infrastructure for ML systems, ensuring data security under HIPAA and SOC2 standards, and collaborating closely with engineers to productionize models. Candidates should have 3+ years of experience in ML Ops and strong proficiency in AWS and Python. This position offers an opportunity to work with cutting-edge AI technologies.

Competenze

  • 3+ years of experience in ML Ops, DevOps, or ML platform engineering.
  • Strong experience with AWS and cloud ecosystems.
  • Proficiency in Python and ML lifecycle tools.

Mansioni

  • Build and maintain scalable model serving infrastructure.
  • Develop CI/CD pipelines for ML models.
  • Integrate observability platforms for ML workflows.

Conoscenze

AWS
ML lifecycle tools
CI/CD
Python
Docker
Kubernetes

Formazione

Bachelor’s or Master’s in Computer Science, Software Engineering, ML/AI

Strumenti

MLflow
Databricks
Terraform
Descrizione del lavoro

Propio is on a mission to make communication accessible to everyone. As a leader in real‑time interpretation and multilingual language services, we connect people with the information they need across language, culture, and modality. We’re committed to building AI‑powered tools to enhance interpreter workflows, automate multilingual insights, and scale communication quality across industries.

Role Summary

The Machine Learning Operations Engineer will design, build, and maintain the production infrastructure required to deploy, scale, monitor, and govern Propio’s ML and agentic AI systems. This role ensures that translation, speech, interpretation, and conversational AI models run reliably, securely, and cost‑effectively in real‑time environments. The MLOps Engineer bridges ML engineering, DevOps, and platform engineering—owning the end‑to‑end operational lifecycle from training pipelines to automated deployment to observability, aligning with HIPAA, SOC2, and HITRUST standards.

Key Responsibilities
  • Build and maintain scalable model serving infrastructure for real‑time inference (translation, ASR/TTS, agentic AI workflows)
  • Implement automated CI/CD pipelines for ML models and LLM agents, including versioning, rollback strategies, and multi‑environment promotion (dev, staging, prod)
  • Develop GPU/compute orchestration strategies for cost‑efficient workloads across AWS (SageMaker, ECS/EKS, EC2, or Databricks)
  • Implement reproducible ML workflows with strong dependency management, data lineage, feature versioning, and reproducibility guarantees
  • Integrate observability platforms (Datadog, MLflow, LangSmith) for end‑to‑end tracing of agentic workflows and multi‑step tool execution
  • Build alerting systems and dashboards for both business‑level metrics (quality, throughput) and engineering metrics (GPU load, memory, queue depth)
  • Ensure ML systems meet HIPAA, SOC2, and HITRUST standards, including encryption, audit logging, access controls, and secure handling of PHI
  • Implement data validation, schema enforcement, and drift detection to guarantee data quality for both training and inference
  • Manage model registry, feature store, and lineage tracking across all AI services
  • Work closely with Machine Learning Engineers to productionize models and agentic systems, ensuring seamless handoff from experimentation to deployment
  • Collaborate with Data Engineering to operationalize data pipelines feeding ML/LLM workflows
  • Partner with DevOps, Security Engineering, and Platform Engineering to integrate ML systems into Propio’s cloud stack
  • Optimize model serving architectures for latency, concurrency, and cost
  • Implement autoscaling, caching, routing, and load‑balancing solutions for high‑volume LLM and speech‑based systems
  • Evaluate and implement new technologies (vector databases, real‑time streaming infra, model compression, quantization)
Requirements
  • 3+ years of experience in ML Ops, DevOps, or ML platform Engineering or similar infrastructure‑focused ML roles
  • Strong experience with AWS (SageMaker, EKS/ECS, Lambda, Step Functions, S3, IAM), Databricks, or equivalent cloud ecosystems
  • Strong proficiency with ML lifecycle tools: MLflow, Kubeflow, SageMaker Pipelines, Airflow, Prefect, or equivalent
  • Strong foundations in CI/CD, containerization (Docker), orchestration (Kubernetes), and infrastructure‑as‑code (Terraform, CloudFormation)
  • Experience implementing monitoring and observability for ML systems (Datadog, Prometheus/Grafana, LangSmith, MLflow)
  • Familiarity with securing ML pipelines and handling regulated data under HIPAA and SOC2
  • Proficiency in Python and experience supporting ML engineers in productionizing ML/LLM workflows
  • Bachelor’s or Master’s in Computer Science, Software Engineering, ML/AI, or related field
Ottieni la revisione del curriculum gratis e riservata.
oppure trascina qui un file PDF, DOC, DOCX, ODT o PAGES di non oltre 5 MB.