Enable job alerts via email!

Data Scientist

E-Solutions

Ottawa

Remote

CAD 100,000 - 130,000

Full time

Today
Be an early applicant

Job summary

A leading tech firm is seeking an MDM Lead Data Engineer with strong expertise in Databricks and Python. The role involves building and optimizing large-scale data pipelines, collaborating with teams to deliver tailored data solutions, and ensuring data governance. The ideal candidate has strong skills in SQL, cloud ETL tools, and experience in Azure services. This position is offered as a remote role with optional certifications valued.

Qualifications

  • Strong expertise in Databricks and related technologies.
  • Experience in big data processing using Spark and PySpark.
  • Certifications like Azure Data Engineer Associate are a plus.

Responsibilities

  • Build and maintain scalable ETL/ELT pipelines using Databricks.
  • Collaborate with cross-functional teams for data requirements.
  • Optimize data pipelines for cost efficiency and performance.
  • Implement data security and governance standards.

Skills

Databricks
Snowflakes
Python
SQL
Datawarehouse
Cloud ETL tools

Tools

Azure Cloud Services
GitLab
Job description
Overview

Role: MDM Lead Data Engineer

Location: Remote (Preferred – Ottawa)

Remote Role

Required: Databricks, Snowflakes, Python

Domain knowledge around reference data, data mgmt. projects

any knowledge of Zennov is a plus

Job Summary

We are seeking highly skilled Azure Data Engineer with strong expertise in SQL, Python, Datawarehouse, Cloud ETL tools to join our data team. The ideal candidate will design, implement and optimize large-scale data pipeline, ensuring scalability, reliability and performance. This role involves working closely with multiple teams and business stakeholders to deliver cutting-edge data solutions.

Key Responsibilities
  • Data Pipeline Development: Build and maintain scalable ETL / ELT pipelines using Databricks. Leverage PySpark / Spark and SQL to transform and process large datasets. Integrate data from multiple sources including Azure Blob Storage, ADLS and other relational / non-relational systems.
  • Collaboration & Analysis: Work closely with multiple teams to prepare data for dashboard and BI Tools. Collaborate with cross-functional teams to understand business requirements and deliver tailored data solutions.
  • Performance & Optimization: Optimize Databricks workloads for cost efficiency and performance. Monitor and troubleshoot data pipelines to ensure reliability and accuracy.
  • Governance & Security: Implement and manage data security, access controls and governance standards using Unity Catalog. Ensure compliance with organizational and regulatory data policies.
  • Deployment: Leverage Databricks Asset Bundles for seamless deployment of Databricks jobs, notebooks and configurations across environments. Manage version control for Databricks artifacts and collaborate with team to maintain development best practices.
Technical Skills
  • Strong expertise in Databricks (Delta Lake, Unity Catalog, Lakehouse Architecture, Table Triggers, Delta Live Pipelines, Databricks Runtime etc.)
  • Proficiency in Azure Cloud Services.
  • Solid Understanding of Spark and PySpark for big data processing.
  • Experience in relational databases.
  • Knowledge on Databricks Asset Bundles and GitLab.
Preferred Experience
  • Familiarity with Databricks Runtimes and advanced configurations.
  • Knowledge of streaming frameworks like Spark Streaming.
  • Experience in developing real-time data solutions.
Certifications
  • Azure Data Engineer Associate or Databricks certified Data Engineer Associate certification. (Optional)
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs