Enable job alerts via email!

Data Engineer

Interfront SOC

Pretoria

Hybrid

ZAR 200 000 - 300 000

Full time

Yesterday
Be an early applicant

Job summary

A technology services company is seeking a Data Engineer to manage data pipelines and optimize data processes. The ideal candidate has significant experience with SQL Server, ETL workflows, and data governance. This hybrid contract role is based in Pretoria, involving collaboration with data teams on innovative data solutions. Apply if you have a strong background in data engineering and a passion for quality data management.

Qualifications

  • 6-8 years of experience in data engineering or database development.
  • Proven hands-on experience with advanced T-SQL development.
  • Experience in building and maintaining data warehouses.

Responsibilities

  • Design, build and optimize T-SQL stored procedures and scripts.
  • Develop, deploy and monitor end-to-end ETL workflows.
  • Implement and monitor data validation and QA frameworks.

Skills

Advanced SQL Server Development
ETL and Data Warehousing
Data Validation and Quality Assurance
Data Governance and Security
Performance Optimization

Education

Tertiary qualification in Computer Science, Information Systems, Data Engineering, Analytics, Mathematics or Statistics

Tools

SQL Server
Azure Data Factory
SSIS
Airflow
Power BI
Job description

Description

Key Responsibilities

Data Engineering & Pipeline Management

Design build and optimize T-SQL stored procedures functions and scripts for high-volume data processing and ECM scoring.

Develop deploy and monitor end-to-end ETL / ELT workflows (e.g. SQL Server Agent SSIS Azure Data Factory or Airflow) with checkpoint / rollback jobtracking and recovery capabilities.

Perform data cleansing preparation and transformation to support business intelligence and machine learning workflows.

Engineer and maintain reusable feature store tables (per entity / tax type) for ML models and operational scoring.

Model and maintain data warehouse structures (3NF dimensional / star / snowflake) ensuring proper documentation of data lineage.

Prepare and deliver curated scored datasets for downstream consumption in Power BI dashboards and analytics environments.

Develop and maintain audit telemetry and job tracking tables to ensure data reliability restartability and monitoring visibility.

Support and troubleshoot production pipelines optimizing query performance via indexing tuning and profiling tools.

Data Quality Governance and Compliance

Implement and monitor data validation reconciliation and QA frameworks across the data lifecycle.

Enforce data security privacy and compliance controls in line with corporate and regulatory standards.

Support the implementation of data governance and lineage documentation ensuring traceability and adherence to EDM policies.

Collaboration and Cross-functional Support

Collaborate with data analysts data scientists software engineers and business stakeholders to translate business problems into scalable data solutions.

Provide accessible well-documented datasets to support analytics and reporting.

Contribute to all phases of the SDLC including requirements design development testing deployment and maintenance.

Closing date : 12th of November 2025

Position type : 6 months contract

Available positions : 2

Hybrid position and successful candidate must be based in Gauteng area

These positions are open to persons with disabilities.

Requirements

Qualifications and Experience

A tertiary qualification in Computer Science Information Systems Data Engineering Analytics Mathematics or Statistics or Matric with 6-8 years ofexperience in data engineering database development or data management in production environments.

Proven hands-on experience with SQL Server including advanced T-SQL development ETL / ELT workflow design and performance tuning.

Demonstrated delivery of production data solutionsboth batch and near real-timewithin enterprise environments.

Experience in building and maintaining data warehouses feature stores and reusable data products.

Track record of implementing data governance and quality frameworks ensuring compliance and traceability.

Experience in orchestrating complex data pipelines using SQL Server Agent SSIS Airflow or Azure Data Factory.

Familiarity with cloud-based data architectures (Azure preferred) and version control systems (Git).

Exposure to Power BI or equivalent visualization tools for reporting and analytics enablement.

Strong understanding of data security privacy and regulatory compliance requirements.

Key Skills and Competencies

Advanced SQL Server Development : Strong proficiency in T-SQL stored procedure design query optimization indexing and error handling.

ETL and Data Warehousing : Expertise in ETL / ELT pipeline design and orchestration for batch and near real-time processing using SQL Server Agent SSIS orAzure Data Factory.

Data Modeling : Solid understanding of normalized and dimensional modeling (3NF star snowflake) and scalable architecture design.

Feature Store Development : Ability to design and maintain reusable feature tables supporting machine learning and operational scoring.

Data Validation and Quality Assurance : Skilled in implementing validation rules reconciliation checks and QA frameworks to ensure data integrity.

Data Governance and Security : Strong knowledge of data governance privacy and compliance standards; experience maintaining data lineage documentation.

Workflow Orchestration : Experience building restartable traceable workflows with checkpoint and rollback mechanisms.

Programming and Scripting : Proficiency in SQL and beneficial experience in Python or R for automation and data manipulation.

Cloud Platforms : Familiarity with Azure (preferred) or other cloud platforms such as AWS or GCP for data engineering workloads.

Version Control and CI / CD : Exposure to Git and CI / CD pipelines for managing data workflow deployment.

Visualization and Reporting (Beneficial) : Ability to prepare scored or curated data for BI tools such as Power BI.

Performance Optimization : Expertise in performance tuning query profiling and indexing strategies to optimize large-scale data operations.

Collaboration and Communication : Ability to work effectively across technical and business teams translating complex requirements into practical datasolutions.

Work Level

Mid-Level

Job Type

Contract

Salary

Market Related

Duration

6 - 9 Months

EE Position

No

Location

Pretoria

Required Experience

Manager

Key Skills

Apache Hive,S3,Hadoop,Redshift,Spark,AWS,Apache Pig,NoSQL,Big Data,Data Warehouse,Kafka,Scala

Employment Type : Contract

Experience : years

Vacancy : 1

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.