Enable job alerts via email!

MLOps Engineer

Koda Staff

Ontario

Remote

CAD 140,000 - 180,000

Full time

12 days ago

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

A leading company is seeking an MLOps guru to optimize and automate machine learning processes. You will design and manage ML pipelines, deploy models on cloud platforms, and ensure real-time performance. Join a team of tech rebels focused on making AI smarter and faster in a flexible work environment.

Benefits

Flexibility to work from anywhere
Competitive salary plus benefits
Learning culture with growth opportunities
Work hard, play hard culture

Qualifications

  • 5+ years of experience in MLOps, DevOps, or AI model deployment.
  • Mastery of Python and ML frameworks like TensorFlow, PyTorch.
  • Hands-on experience with Docker, Kubernetes, and serverless architectures.

Responsibilities

  • Design, develop, and optimize ML pipelines ensuring smooth operations.
  • Automate deployment of deep learning and generative AI models.
  • Implement model versioning and manage ML models on cloud platforms.

Skills

Python
MLOps
TensorFlow
PyTorch
Docker
Kubernetes
CI/CD

Education

BS or MS in Computer Science

Tools

Docker
Kubernetes
AWS
GCP
Azure

Job description

1 month ago Be among the first 25 applicants

This range is provided by Koda Staff. Your actual pay will be based on your skills and experience — talk with your recruiter to learn more.

Base pay range

$140,000.00/yr - $180,000.00/yr

Direct message the job poster from Koda Staff

Helping clients find top USA based engineers across Blockchain, Generative AI, and more! Need Golang, Python, Rust, or React gurus in the space? Look…

On a mission to make AI smarter, faster, and everywhere. If you live and breathe machine learning pipelines, cloud platforms, and deployment automation, we want you to join my client's team of tech rebels.

They’re not looking for a mere mortals— we want an MLOps guru ready to optimize, automate, and scale machine learning to the moon.

What you’ll be doing:

  • Design, develop, and optimize ML pipelines like a wizard—ensuring training, validation, and inference run smoothly at scale.
  • Automate deployment of deep learning and generative AI models for real-time applications (because who wants to deploy manually?).
  • Implement model versioning, rollbacks, and ensure seamless updates like a smooth operator.
  • Deploy and manage ML models on cloud platforms (AWS, GCP, Azure) using your containerized magic (hello Docker and Kubernetes).
  • Optimize real-time inference performance—bring TensorRT, ONNX, and PyTorch to their full glory.
  • Work with GPU acceleration, distributed computing, and parallel processing to make AI workloads faster than a rocket.
  • Fine-tune models to slice latency and boost scalability—because who likes slow models?
  • Build and maintain CI/CD pipelines for ML models (GitHub Actions, Jenkins, ArgoCD) to make life easier.
  • Automate retraining and deployment to ensure the AI is always learning (and never slacking).
  • Develop monitoring solutions to track model drift, data integrity, and performance, because we don’t believe in letting things slide.
  • Stay on top of security, data privacy, and AI ethics standards—because we care about doing things right.

What we need from you:

  • 5+ years of experience in MLOps, DevOps, or AI model deployment (you’ve been around the block).
  • Mastery of Python and ML frameworks like TensorFlow, PyTorch, and ONNX (you know these like the back of your hand).
  • You’ve deployed models using Docker, Kubernetes, and serverless architectures—it’s second nature to you.
  • Hands-on experience with ML pipeline tools (ArgoWorkflow, Kubeflow, MLflow, Airflow)—you’ve built some mean pipelines.
  • Expertise in cloud platforms (AWS, GCP, Azure) and hosting AI/ML models like a pro.
  • GPU-based inference acceleration experience (CUDA, TensorRT, NVIDIA DeepStream)—you make inference fast!
  • Solid background in CI/CD workflows, automated testing, and deploying ML models without a hitch.
  • Real-time inference optimization? Check. Scalable ML infrastructure? Double check.
  • Excellent technical judgment—you can architect a system that works today and evolves tomorrow.
  • You’re all about automation—if you can script it, you will.
  • You’ve got a deep understanding of distributed systems and computing architectures.
  • Self-driven with the ability to work independently and own it.
  • Experience with Kubernetes, Docker, or microservices in general—this is your bread and butter.
  • BS or MS in Computer Science or equivalent—because you’ve got the education (or experience) to back it up.

Nice to have:

  • Some CUDA programming skills (you’ve probably dabbled).
  • Experience with LLMs and generative AI models in production.
  • A bit of networking knowledge never hurt anyone.
  • Familiarity with distributed computing frameworks (Ray, Horovod, Spark)—you like to go big.
  • Edge AI deployment experience (Triton Inference Server, TFLite, CoreML)—because why not push the envelope?

Why Them?

  • Work with top-tier AI experts in a fast-growing startup.
  • Flexibility—work from anywhere, anytime, as long as you get stuff done.
  • Competitive salary plus benefits (obviously).
  • Learning culture—we provide opportunities to grow and expand your skills.
  • A work hard, play hard culture—because who says you can’t have fun while crushing it?
Seniority level
  • Seniority level
    Mid-Senior level
Employment type
  • Employment type
    Full-time
Job function
  • Job function
    Information Technology
  • Industries
    Software Development

Referrals increase your chances of interviewing at Koda Staff by 2x

Sign in to set job alerts for “Machine Learning Engineer” roles.
Machine Learning Developer Intern (4 month term - Toronto or Montreal)
Machine Learning, Optimization. Remote or Hybrid
Data Scientist Lead, AI and Data - Elevate Program 2025

Toronto, Ontario, Canada CA$150,000.00-CA$240,000.00 1 month ago

Software Engineer I, Entry Level (Fall 2024-Spring 2025) - Toronto
Software Engineer, Backend (All Levels / All Teams)
Machine Learning, Software Engineer (Remote)

Toronto, Ontario, Canada $190,000.00-$240,000.00 1 month ago

Machine Learning Engineer, Support Experience
Software Engineer, Machine Learning Infrastructure

Greater Toronto Area, Canada CA$120,000.00-CA$150,000.00 5 days ago

We’re unlocking community knowledge in a new way. Experts add insights directly into each article, started with the help of AI.

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Principal MLOPs Engineer (Canada)

Rackspace Technology

Toronto

Remote

CAD 120,000 - 175,000

4 days ago
Be an early applicant

Principal MLOPs Engineer (Canada)

Rackspace Technology

Toronto

Remote

CAD 120,000 - 165,000

5 days ago
Be an early applicant

Principal MLOPs Engineer (Canada)

Rackspace Technology

Toronto

Remote

CAD 125,000 - 150,000

30+ days ago

MLOps Engineer

Manulife Insurance Malaysia

Toronto

Hybrid

CAD 75,000 - 141,000

8 days ago

MLOps Engineer

Manulife

Waterloo

Hybrid

CAD 75,000 - 141,000

26 days ago

MLOps Engineer

Manulife

Toronto

On-site

CAD 75,000 - 141,000

26 days ago

Senior MLOps Developer

Elastic

Remote

CAD 128,000 - 203,000

29 days ago

Senior MLOps Developer

Referral Board

Remote

CAD 128,000 - 203,000

30+ days ago

Senior MLOps Developer

Elasticsearch B.V.

Remote

CAD 128,000 - 203,000

30+ days ago