Enable job alerts via email!

Data Engineer

Talent Bridge HR Consultancy Dubai

Dubai

On-site

AED 120,000 - 200,000

Full time

Yesterday
Be an early applicant

Job summary

A leading consultancy in Dubai is seeking an experienced professional to build robust data pipelines that power research and production AI. The role involves architecting batch/stream ETL/ELT pipelines, managing schemas, and ensuring data quality and governance. Candidates should have strong programming skills in C or Java, cloud experience, and familiarity with MLOps practices. This position offers an opportunity to work on innovative AI solutions.

Qualifications

  • Solid Linux background for development and deployment.
  • Experience with model packaging and deployment fundamentals.

Responsibilities

  • Architect and operate batch/stream pipelines for ETL/ELT.
  • Model/manage schemas and enforce data quality.
  • Support ML workflows with DVC and MLflow or W&B.
  • Build feature stores/data services and expose datasets.
  • Optimize performance/cost across storage/compute.

Skills

Programming in C or Java
SQL & NoSQL
Pandas/NumPy
PySpark
Airflow
API development
Docker

Education

Proven experience building reliable pipelines in production

Tools

DVC
MLflow
AWS SageMaker
Azure ML
GCP AI
Job description
Role Summary

Build robust observable data pipelines that power research and production AI. Success means high pipeline reliability (on-time SLAs), strong data quality (validation & lineage) and enabling fast experimentation. You will partner with AI/ML analytics and product to make data trustworthy and usable.

Responsibilities
  • Architect and operate batch/stream pipelines (Airflow; Spark optional) for ETL/ELT.
  • Model/manage schemas; enforce data quality and lineage/governance.
  • Support ML workflows with DVC (data versioning) and MLflow or Weights & Biases.
  • Build feature stores/data services; expose datasets via secure REST endpoints.
  • Optimize performance/cost across storage/compute; implement monitoring/alerting.
  • Maintain documentation and internal catalogs; enable self-service analytics.
Qualifications
  • Skills: Programming in C or Java; SQL & NoSQL; Pandas/NumPy; PySpark; Airflow; API development; Docker.
  • MLOps: DVC; MLflow or W&B; model packaging/deployment fundamentals.
  • Cloud: AWS SageMaker Azure ML or GCP AI experience.
  • Nice to have: Unreal Engine exposure.
  • Environment: Solid Linux background for development and deployment.
  • Education/Experience: Proven experience building reliable pipelines in production.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.