Enable job alerts via email!

AI Data Engineer

Discovered MENA

Dubai

On-site

AED 120,000 - 200,000

Full time

2 days ago
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Start fresh or import an existing resume

Job summary

A data engineering firm in Dubai seeks an AI Data Engineer to design and maintain data pipelines, ensuring data flow and accessibility for AI initiatives. The ideal candidate will have 3-5 years of experience in data engineering, working with on-premise and cloud platforms like AWS and GCP. A strong programming background and familiarity with big data technologies are essential. This role is fully on-site in Dubai.

Qualifications

  • Minimum of 3-5 years of experience in data engineering or a similar role.
  • Proven experience with on-premise and cloud platforms (AWS, GCP, Azure).
  • Strong background in data integration, ETL processes, and data pipeline development.

Responsibilities

  • Design, develop, and maintain robust and scalable data pipelines.
  • Integrate data from various internal and external sources.
  • Develop and implement ETL processes to automate data ingestion.

Skills

Proficiency in scripting and programming languages (e.g., Python, SQL, Bash)
Strong knowledge of data storage solutions and databases (e.g., SQL, NoSQL, data lakes)
Experience with big data technologies (e.g., Apache Spark, Hadoop)
Experience with CI/CD tools (e.g., Jenkins, GitLab CI, CircleCI)
Understanding of data engineering and MLOps methodologies
Awareness of security best practices in data environments

Tools

AWS
GCP
Azure

Job description

Location: Fully on site in Dubai

Rate: 20k - 33k

The AI Data Engineer will play a crucial role in designing, building, and maintaining scalable data infrastructure and pipelines. This position involves working closely with data scientists, AI engineers, and software developers to ensure efficient data flow and accessibility for our AI and data initiatives.

Key Responsibilities:

Infrastructure Management:

● Design, develop, and maintain robust and scalable data pipelines to handle large

datasets using both on-premise and cloud platforms (e.g., AWS, GCP, Azure).

● Implement and manage data storage solutions, including databases and data lakes,

ensuring data integrity and performance.

Data Integration:

● Integrate data from various internal and external sources such as databases, APIs, flat

files, and streaming data.

● Ensure data consistency, quality, and reliability through rigorous validation and

transformation processes.

ETL Development:

● Develop and implement ETL (Extract, Transform, Load) processes to automate data

ingestion, transformation, and loading into data warehouses and lakes.

● Optimize ETL workflows to ensure efficient processing and minimize data latency.

Data Quality & Governance:

● Implement data quality checks and validation processes to ensure data accuracy and

completeness.

● Develop data governance frameworks and policies to manage data lifecycle, metadata,

and lineage.

Collaboration and Support:

● Work closely with data scientists, AI engineers, and developers to understand their data

needs and provide technical support.

● Facilitate effective communication and collaboration between the AI and data teams and

other technical teams.

Continuous Improvement:

● Identify areas for improvement in data infrastructure and pipeline processes.

● Stay updated with the latest industry trends and technologies related to data engineering

and big data.

Experience:

● Minimum of 3-5 years of experience in data engineering or a similar role.

● Proven experience with on-premise and cloud platforms (AWS, GCP, Azure).

● Strong background in data integration, ETL processes, and data pipeline development.

● Led the design and development of high-performance AI and data platforms,

including IDEs, permission management, data pipelines, code management and

model deployment systems.

Skills:

● Proficiency in scripting and programming languages (e.g., Python, SQL, Bash).

● Strong knowledge of data storage solutions and databases (e.g., SQL, NoSQL, data

lakes).

● Experience with big data technologies (e.g., Apache Spark, Hadoop).

● Experience with CI/CD tools (e.g., Jenkins, GitLab CI, CircleCI).

● Understanding of data engineering and MLOps methodologies.

● Awareness of security best practices in data environments.

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.