Job Search and Career Advice Platform

Enable job alerts via email!

Data Engineer II

LexisNexis

Wes-Kaap

On-site

ZAR 300 000 - 400 000

Full time

Today
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Job summary

A leading information services firm in South Africa is seeking a Data Engineer to contribute to building the next-generation Strategic Data Platform. You will design and maintain data pipelines, collaborate with senior engineers, and ensure high-quality data delivery. The ideal candidate will have hands-on experience in data engineering, proficiency in Python, and a strong willingness to learn. This role offers an innovative work environment with strong support for professional development and well-being initiatives.

Benefits

Medical Aid
Retirement Plan with Risk Benefits
Study Leave
Flexible Working Hours

Qualifications

  • Hands-on experience in a software/data engineering role.
  • Proficiency in Python and familiarity with PySpark or similar frameworks.
  • Understanding of data delivery patterns, including REST APIs.

Responsibilities

  • Develop and maintain data pipelines using Python and PySpark.
  • Deliver enriched datasets to customers via Databricks and APIs.
  • Collaborate with cross-functional teams including engineers and product managers.

Skills

Python
PySpark
Data Analysis
Collaboration
Problem-Solving
Communication

Tools

Databricks
SQL
Git
Job description

Job Description

Would you like to ensure the successful delivery of the Data Platform and Software Innovations?

Do you enjoy creating a collaborative and customer-focused working environment?

About Team:

LexisNexis Intellectual Property, which serves customers in more than 150 countries with 11,300 employees worldwide, is part of RELX, a global provider of information-based analytics and decision tools for professional and business customers.

About Role:

As a Data Engineer at LexisNexis Intellectual Property (LNIP), you’ll contribute to building and maintaining our next-generation Strategic Data Platform. This platform ingests, enriches, and transforms global patent and IP-related data to power key products like PatentSight+, as well as a growing ecosystem of internal tools and customer-facing solutions.

In this early-career role, you will collaborate with senior engineers and technical leads to design robust data pipelines, apply engineering best practices, and support the delivery of high‑quality data through modern platforms such as Databricks, APIs, and event‑driven systems. You'll gain practical experience working at scale while contributing to the delivery of data directly to customers and systems across the organisation.

Key Responsibilities:
  • Contributing to the development and maintenance of data pipelines using Python, PySpark, and Databricks
  • Supporting the delivery of enriched datasets to customers via Databricks, RESTful APIs, and event‑driven delivery mechanisms (e.g., Kafka or similar)
  • Assisting in data ingestion, transformation, and enrichment across the medallion architecture (bronze → silver → gold)
  • Collaborating with cross‑functional teams, including engineers, data analysts, and product managers
  • Participating in code reviews, unit testing, and documentation to ensure high code quality and maintainability
  • Troubleshooting and debugging data issues across development and production environments
  • Following and contributing to internal best practices around data engineering and software development
  • Continuously developing technical skills and understanding of the business domain
Requirements:
  • Hands‑on experience in a software/data engineering role.
  • Proficiency in Python and working knowledge of PySpark or similar distributed data frameworks.
  • Familiarity with Databricks or a strong interest in learning and working with the platform.
  • Understanding of data delivery patterns, including REST APIs and event‑driven architectures.
  • Experience with SQL and structured data manipulation.
  • Familiarity with version control systems (e.g., Git).
  • Strong problem‑solving mindset and willingness to learn from feedback.
  • Good communication skills and ability to work in a team setting.
Nice to Have:
  • Exposure to cloud platforms like AWS, Azure, or GCP.
  • Experience working with large‑scale or open datasets.
  • Familiarity with medallion architecture or similar data lake patterns.
  • Understanding of data quality principles and CI/CD pipelines for data workflows.
Why Join Us?

Join our team and contribute to a culture of innovation, collaboration, and excellence. If you are ready to advance your career and make a significant impact, we encourage you to apply.

Work in a way that works for you

We promote a healthy work/life balance across the organisation. We offer an appealing working prospect for our people. With numerous wellbeing initiatives, shared parental leave, study assistance and sabbaticals, we will help you meet your immediate responsibilities and your long‑term goals.

  • Working flexible hours – flexing the times when you work in the day to help you fit everything in and work when you are the most productive
Working for you

We know that your well‑being and happiness are key to a long and successful career. These are some of the benefits we are delighted to offer:

  • Medical Aid
  • Retirement Plan inclusive of Risk Benefits (Disability, Critical Illness, Life Cover & Funeral Cover)
  • Modern family benefits, including adoption and surrogacy
  • Study Leave
About the Business

LexisNexis Legal & Professional® provides legal, regulatory, and business information and analytics that help customers increase their productivity, improve decision‑making, achieve better outcomes, and advance the rule of law around the world. As a digital pioneer, the company was the first to bring legal and business information online with its Lexis® and Nexis® services.

Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.