Job Search and Career Advice Platform

Enable job alerts via email!

Data Engineer - 0122

Jordan Hr

Gauteng

On-site

ZAR 600 000 - 800 000

Full time

4 days ago
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Job summary

A data solutions provider in Gauteng is seeking a Data Engineer to design and maintain data pipelines on Databricks. The ideal candidate will have expertise in Apache Spark, Python, and cloud platforms like Azure. With at least 3 years of experience, you will collaborate with teams to optimize data workflows and ensure data quality and compliance. This role offers the opportunity to work on cutting-edge data solutions in a fast-paced environment.

Qualifications

  • Bachelor's or Master's in Computer Science, Data Engineering, or related field.
  • 3+ years of hands-on data engineering experience.

Responsibilities

  • Design, develop, and maintain robust data pipelines on Databricks.
  • Architect scalable data models for reporting and analytics.
  • Monitor data quality metrics and resolve issues.
  • Collaborate to optimize performance and cost management in Azure Databricks.

Skills

Apache Spark (PySpark)
Databricks notebooks
SQL
Python programming
Azure cloud platforms
ETL / ELT pipelines
Data governance
Problem-solving skills

Education

Bachelor's or Master's in Computer Science, Data Engineering, Information Systems

Tools

Databricks
Delta Lake
GitHub Actions
Job description

Our client is seeking a hands‑on Data Engineer with strong experience in building scalable data pipelines and analytics solutions on Databricks.

They will design, implement, and maintain end‑to‑end data flows, optimize performance, and collaborate with data scientists, analytics, and business stakeholders to turn raw data into trusted insights.

ESSENTIAL SKILLS
  • Expertise with Apache Spark (PySpark), Databricks notebooks, Delta Lake, and SQL
  • Strong programming skills in Python for data processing
  • Experience with cloud data platforms (Azure) and their Databricks offerings; familiarity with object storage (ADLS)
  • Proficient in building and maintaining ETL / ELT pipelines, data modeling, and performance optimization
  • Knowledge of data governance, data quality, and data lineage concepts
  • Experience with CI / CD for data pipelines, and orchestration tools (GitHub Actions, Asset Bundles or Databricks jobs)
  • Strong problem‑solving skills, attention to detail, and ability to work in a collaborative, cross‑functional team
ADVANTAGEOUS SKILLS
  • Experience with streaming data (Structured Streaming, Kafka, Delta Live Tables)
  • Familiarity with materialized views, streaming tables, data catalogs and metadata management
  • Knowledge of data visualization and BI tools (Splunk, Power BI, Grafana)
  • Experience with data security frameworks and compliance standards relevant to the industry
QUALIFICATIONS / EXPERIENCE
  • Bachelors or Masters degree in Computer Science, Data Engineering, Information Systems, or a related field.
  • 3+ years of hands‑on data engineering experience.

Certifications in Databricks or cloud provider platforms.

Key Responsibilities
  • Design, develop, test, and maintain robust data pipelines and ETL / ELT processes on Databricks (Delta Lake, Spark, SQL, Python / Scala / SQL notebooks)
  • Architect scalable data models and data vault / dimensional schemas to support reporting, BI, and advanced analytics
  • Implement data quality, lineage, and governance practices; monitor data quality metrics and resolve data issues proactively
  • Collaborate with Data Platform Engineers to optimize cluster configuration, performance tuning, and cost management in cloud environments (Azure Databricks)
  • Build and maintain data ingestion from multiple sources (RDBMS, SaaS apps, files, streaming queues) using modern data engineering patterns (CDC, event‑driven pipelines, change streams, Lakeflow Declarative Pipelines)
  • Ensure data security and compliance (encryption, access controls) in all data pipelines
  • Develop and maintain CI / CD pipelines for data workflows; implement versioning, testing, and automated deployments
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.