Enable job alerts via email!

Data Engineer

Alltech Consulting Services

Mississauga

Hybrid

CAD 100,000 - 130,000

Full time

Today
Be an early applicant

Generate a tailored resume in minutes

Land an interview and earn more. Learn more

Job summary

A consulting firm specializing in data solutions is seeking a Cloud Data Engineer to build and optimize data pipelines, design and deploy Spark programs in Databricks, and ensure data integrity. Candidates should have over 5 years of experience in Python and SQL, along with robust knowledge in ETL processes and cloud data platforms like Snowflake and Synapse. This role is critical in managing and optimizing enterprise-level cloud data solutions.

Qualifications

  • 5+ years experience in Python coding.
  • 5+ years in SQL Server based development of large datasets.
  • 5+ years with ETL pipeline development using Databricks Pyspark.
  • Experience with cloud data warehouses and architectures.

Responsibilities

  • Build and optimize data pipelines for data ingestion and transformation.
  • Develop Spark programs in Databricks to process large data volumes.
  • Monitor and troubleshoot Databricks jobs.
  • Implement data management best practices.

Skills

Python
SQL
Pyspark
Data Warehouse Concepts
ETL Pipelines
Databricks
Cloud Data Platforms
Dimensional data modeling
Airflow

Tools

SQL Server
Synapse
Big Query
Redshift
Snowflake
Job description

This position is for a Cloud Data engineer with a background in Python, Pyspark, SQL and data warehousing for enterprise level systems. The position calls for someone that is comfortable working with business users along with business analyst expertise.

Major Responsibilities
  • Build and optimize data pipelines for efficient data ingestion, transformation and loading from various sources while ensuring data quality and integrity.
  • Design, develop, and deploy Spark program in databricks environment to process and analyze large volumes of data.
  • Experience of Delta Lake, DWH, Data Integration, Cloud, Design and Data Modelling.
  • Proficient in developing programs in Python and SQL
  • Experience with Data warehouse Dimensional data modeling.
  • Working with event based / streaming technologies to ingest and process data.
  • Working with structured, semi structured and unstructured data.
  • Optimize Databricks jobs for performance and scalability to handle big data workloads.
  • Monitor and troubleshoot Databricks jobs, identify and resolve issues or bottlenecks.
  • Implement best practices for data management, security, and governance within the Databricks environment. Experience designing and developing Enterprise Data Warehouse solutions.
  • Proficient writing SQL queries and programming including stored procedures and reverse engineering existing process.
  • Perform code reviews to ensure fit to requirements, optimal execution patterns and adherence to established standards.
Skills
  • 5+ years Python coding experience.
  • 5+ years – SQL Server based development of large datasets
  • 5+ years with Experience with developing and deploying ETL pipelines using Databricks Pyspark.
  • Experience in any cloud data warehouse like Synapse, Big Query, Redshift, Snowflake.
  • Experience in Data warehousing – OLTP, OLAP, Dimensions, Facts, and Data modeling.
  • Previous experience leading an enterprise-wide Cloud Data Platform migration with strong architectural and design skills.
  • Experience with Cloud based data architectures, messaging, and analytics.
  • Cloud certification(s).
  • Any experience with Airflow is a Plus.
Must Have
  • Snowflake
  • SQL
  • Python
  • ADF
  • PySpark
  • Data Warehouse Concepts
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.