Enable job alerts via email!

Machine Learning Operation Engineer (MLOps Engineer)

Always Marketing Malaysia Sdn Bhd

Kuala Lumpur

On-site

MYR 60,000 - 90,000

Full time

13 days ago

Job summary

A technology solutions company in Kuala Lumpur seeks a cloud infrastructure specialist to work closely with data scientists and engineers on ML model deployment. The ideal candidate has a degree in computer science or engineering and over 2 years' experience in managing cloud infrastructure and Kubernetes. Responsibilities include designing scalable infrastructure, developing APIs, and ensuring seamless integration across various platforms.

Qualifications

  • Minimum 2 years of hands-on experience managing cloud infrastructure in a production environment.
  • Hands-on experience with Kubernetes for container orchestration, scaling and deployment of ML services.
  • Proficient in building and maintaining RESTful or gRPC APIs for ML inference and data services.
  • Proven experience working with relational databases such as Microsoft SQL Server and PostgreSQL.
  • Demonstrated expertise in building scalable and maintainable API services using Python frameworks.
  • Experience in integrating models using NCNN, MNN, or ONNX Runtime Mobile on mobile and edge devices.
  • Experience with Docker/Podman, CI/CD pipelines, Git, and ML lifecycle tools.
  • Strong analytical and troubleshooting skills.

Responsibilities

  • Provide deep technical expertise on cloud infrastructure design and API development.
  • Work closely with data scientists and software engineers for ML models deployment.
  • Set up and maintain cloud and edge infrastructure for ML deployments.
  • Design, implement, and maintain scalable infrastructure for ML workloads.
  • Ensure solutions are well designed with maintainability across platforms.

Skills

Cloud Infrastructure & Kubernetes
API Development & Messaging Integration
System Design, Database & Software Architecture
Programming Languages
Edge AI Deployment
MLOps & Tooling
Monitoring & Logging
Soft Skills & Collaboration

Education

Bachelor's degree in computer science, Engineering or related subject

Tools

AWS
GCP
Azure
Kubernetes
Flask
FastAPI
Docker
Prometheus
Grafana
ELK stack

Job description

Duties and Responsibilities:
  • Provides deep technical expertisein the aspects of cloud infrastructuredesign and APIdevelopment for the business environments.
  • Bridgesthe gap between data scientists and software engineers, enabling the efficient and reliable delivery of ML - powered solutions
  • Ensures solutions are well designedwith maintainability/ease of integration and testing across multipleplatforms.
  • Possess strong proficiency in development and testing practices common to the industry
Summary of Principal Job Responsibility & Specific Job Duties and Responsibilities:
  • Working closely with data scientists, ML engineers, and other stakeholders to deploy ML models
  • Setting up and maintaining cloud and edge infrastructure for MIL models deployment
  • Design, implement and maintain scalable infrastructure for ML workloads
  • Good verbal and written communication skills
  • Collaborative and oriented
Academic Qualification (s):
Bachelor's degree in computer science, Engineering or related subject and/or equivalent formal training or work experience

Work Experience / Skills Requirement(s):

1. Cloud Infrastructure & Kubernetes
  • Minimum 2 years of hands-on experience managing cloud infrastructure (e.g. AWS,GCP,Azure) in a production environment
  • Hands-on experience with Kubernetes for container orchestration, scaling and deployment of ML services
  • Familiar with Helm charts, ConfigMaps, Secret and autoscaling strategies
2. API Development & Messaging Integration
  • Proficient in building and maintaining RESTful or gRPC APIs for ML inference and data services
  • Experience in message queue integration such as RabbitMQ or ZeroMQ for asyncronous communication, job queuing or real-time model inference pipelines
3. System Design, Database & Software Architecture
  • Proven experience working with relational databases (RDBMS) such as Microsoft SQL Server and PostgreSQL.
  • Proficient in schema design, writing complex queries, stored procedures, indexing strategies, and query optimization.
  • Hands-on experience with vector search and embedding-based retrieval systems.
  • Practical knowledge using FAISS, LanceDB, or Qdrant for building similarity search or semantic search pipelines.
  • Understanding of vector indexing strategies (e.g., HNSW, IVF), embedding dimensionality management, and integration with model inference pipelines.
4. Programming Languages
  • Demonstrated expertise in building scalable and maintainable API services using Python frameworks such as Flask, FastAPI, or Litestar.
  • Fluent in HTML, CSS, and JavaScript for building simple web-based dashboards and monitoring interfaces.
  • Experience with Go, C++, or Rust is a strong plus, especially for performance-critical or low-latency inference applications.
5. Edge AI Deployment
  • Experience in integrating models using NCNN, MNN, or ONNX Runtime Mobile on mobile and edge devices.
  • Familiarity with quantization, model optimization, and mobile inference profiling tools.
6. MLOps & Tooling
  • Experience with Docker/Podman, CI/CD pipelines, Git, and ML lifecycle tools such as MLflow, Airflow, or Kubeflow.
  • Exposure to model versioning, A/B testing, and automated re-training workflows.
7. Monitoring & Logging
  • Ability to set up monitoring (e.g., Prometheus, Grafana) and logging (e.g., ELK stack, Loki) to track model performance and system health.
8. Soft Skills & Collaboration
  • Strong analytical and troubleshooting skills.
  • Able to work closely with data scientists, backend engineers, and DevOps to deploy and maintain reliable ML systems.
  • Excellent communication and documentation habits.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.