Enable job alerts via email!

Data Scientist – Agentic RAG & LLM (Databricks / Azure / AWS) - Australia & New Zealand

Rhino Partners

Singapore

On-site

SGD 129,000 - 181,000

Full time

29 days ago

Job summary

A technology consultancy is seeking a skilled Data Scientist to design and produce AI/ML systems using Databricks, Azure, and AWS. The ideal candidate will have expertise in building scalable deployments and CI/CD pipelines. Responsibilities include developing AI workflows and ensuring data governance. Strong background in ML/AI solutions is essential, along with experience in cloud environments.

Qualifications

  • 4+ years of professional experience delivering ML/AI or data science solutions.
  • Strong expertise with the Databricks ecosystem.
  • Hands-on experience with cloud-native ML frameworks.

Responsibilities

  • Design and implement Agentic RAG pipelines using Databricks.
  • Develop agent-based workflows using LangChain and other frameworks.
  • Build CI/CD pipelines for ML & GenAI workloads.

Skills

Databricks
Azure
AWS
Python
SQL
MLOps best practices
CI/CD pipelines
Agentic RAG pipelines
LLM orchestration frameworks

Education

Bachelor’s or Master’s degree in Data Science, Computer Science, AI/ML

Tools

Azure DevOps
GitHub Actions
Jenkins
Terraform
Docker
Job description

Location: Australia & New Zealand (candidates must have valid working rights in either country)

Position Overview

We are seeking a highly skilled Data Scientist with strong expertise in Databricks, Azure, and AWS, specializing in Agentic Retrieval-Augmented Generation (RAG) and Large Language Models (LLMs). The role focuses on designing and productionizing intelligent AI/ML systems with scalable, cloud-native deployments, CI/CD pipelines, and MLOps best practices.

The ideal candidate is hands-on, solution-oriented, and experienced in building and deploying advanced AI systems across multiple cloud platforms.

Key Responsibilities
  • Design and implement Agentic RAG pipelines using Databricks Vector Search, MLflow, Unity Catalog, integrated with Azure Cognitive Search and AWS OpenSearch.
  • Develop agent-based workflows using LangChain, LangGraph, LlamaIndex, and other tool-augmented reasoning frameworks.
  • Fine-tune, evaluate, and deploy LLMs (OpenAI, Anthrop ic, MosaicML, Hugging Face, Llama) for enterprise applications.
  • Build CI/CD pipelines for ML & GenAI workloads, including:
    • Automated build/test/deploy workflows (Azure DevOps, GitHub Actions, Jenkins, AWS CodePipeline).
    • MLflow model registry integration with production/staging environments.
    • Infrastructure-as-Code (IaC) using Terraform, Bicep, or CloudFormation for reproducible deployments.
  • Implement MLOps best practices: experiment tracking, versioning, continuous evaluation, automated retraining pipelines.
  • Ensure data governance, compliance, and security for sensitive datasets across Azure and AWS.
  • Collaborate with engineering and product teams to integrate ETL/ELT pipelines in Azure Data Factory, Synapse, AWS S3, Redshift, Glue.
  • Deploy and monitor models with online evaluation pipelines (MLflow Evaluate, DeepEval, custom scorers such as faithfulness, retrieval recall).
  • Provide technical mentorship on GenAI architecture, CI/CD, and production-grade LLM deployments.
Required Skills & Qualifications
  • Bachelor’s or Master’s degree in Data Science, Computer Science, AI/ML, or related fields (PhD optional, not mandatory).
  • 4+ years of professional experience delivering ML/AI or data science solutions, including cloud-native deployments.
  • Strong expertise with the Databricks ecosystem: Spark (PySpark/Scala), Delta Lake, Unity Catalog, MLflow, Vector Search.
  • Hands-on experience with CI/CD pipelines for ML and GenAI:
    • Azure DevOps, GitHub Actions, or Jenkins.
    • Automated testing for ML pipelines.
    • Model promotion workflows (dev → staging → prod).
  • Proficiency inPython, SQL, distributed data processing, and cloud-native ML frameworks.
  • Deep experience withAzure ML, Data Factory, Synapse, Data Lake andAWS SageMaker, Glue, S3, Redshift.
  • Strong knowledge of LLM orchestration frameworks (LangChain, LangGraph, LlamaIndex).
  • Solid understanding of LLM & RAG evaluation metrics (faithfulness, token-F1, citation@k).
  • Must have valid working rights in Australia or New Zealand.
Preferred Qualifications
  • Experience deploying multi-agent LLM systems in production.
  • Familiarity with Infrastructure-as-Code (Terraform, Bicep, CloudFormation) for CI/CD automation.
  • Hands-on experience with containerization and orchestration (Docker, Kubernetes, AKS, EKS).
  • Contributions to open-source GenAI/LLM projects or published research.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.