At DVT, a leading global custom software development and data engineering company, we are seeking passionate Java developers to join our innovative and collaborative team! This is an opportunity to work alongside some of the most skilled professionals in the industry, leveraging cutting-edge technologies and best practices to deliver world-class solutions.
At DVT, you’ll be part of a culture that fosters continuous learning and growth. We support your professional development with comprehensive training programs and sponsor various industry events like DevConf and GDG. Join us and push the boundaries of what’s possible, while taking your career to new heights!
Position: MLOps Engineer
We are looking for an experienced MLOps Engineer to join our team. The ideal candidate will have a strong background in machine learning, cloud infrastructure, and DevOps practices. The MLOps Engineer will collaborate with data scientists, machine learning engineers, and software developers to deploy, scale, monitor, and manage machine learning models in production. This role requires a blend of software engineering, data science, and operations knowledge.
Key Responsibilities - MLOps:
- Lead the MLOps charge
- Collaborate with data scientists and machine learning engineers to deploy models in production environments using CI / CD pipelines.
- Build and maintain infrastructure for machine learning model training, validation, deployment, and monitoring.
- Implement best practices for data versioning, model versioning, and continuous integration / continuous deployment (CI / CD) pipelines.
- Optimize and automate workflows for machine learning pipelines, ensuring scalability, efficiency, and reproducibility.
- Monitor machine learning models in production for performance drift, accuracy, and degradation.
- Work with cloud platforms (e.g., AWS, GCP, Azure) to set up secure and scalable environments for machine learning workloads.
- Implement monitoring tools and logging solutions for ML model performance, ensuring quick identification and resolution of production issues.
- Develop automated testing and validation scripts to ensure the integrity and reliability of models in production.
- Collaborate with cross-functional teams to improve the end-to-end lifecycle of machine learning projects.
- Ensure compliance with security and data privacy standards throughout the machine learning lifecycle.
Knowledge and Skills:
- Bachelor’s or Master’s degree in Computer Science, Data Science, Machine Learning, Engineering, or a related field.
- 3+ years of experience in MLOps, DevOps, or Data Engineering.
- Strong experience with cloud platforms such as AWS, Google Cloud, or Azure.
- Proficiency in containerization and orchestration tools (Docker, Kubernetes).
- Experience with ML frameworks like TensorFlow, PyTorch, or Scikit-learn.
- Familiarity with version control tools (Git, DVC) and CI / CD pipelines (Jenkins, GitLab CI).
- Experience with infrastructure-as-code tools (Terraform, CloudFormation).
- Proficiency in scripting languages such as Python, Bash, or Scala.
- Experience in monitoring tools (Prometheus, Grafana, ELK Stack) for model and infrastructure health checks.
- Understanding of model training, validation, and deployment processes.
Preferred:
- Experience with automated model retraining and optimization.
- Familiarity with data engineering workflows including ETL, data pipelines, and data governance.
- Experience with feature stores and model repositories.
- Knowledge of A / B testing and model performance tracking (MLflow, TFX, SageMaker, or similar).
- Certifications in cloud computing (AWS, GCP, or Azure).
Key Skills:
- Strong problem-solving skills and ability to work independently.
- Excellent communication skills for cross-functional collaboration.
- Strong understanding of data governance, security, and privacy issues.
- Ability to work in an agile, fast-paced environment.