Enable job alerts via email!

PySpark Developer

Cloudious

Toronto

On-site

CAD 80,000 - 100,000

Full time

Today
Be an early applicant

Job summary

A data solutions company in Toronto is seeking a skilled SAS to PySpark Developer to lead the migration of legacy SAS-based analytics workflows to scalable PySpark solutions. The ideal candidate will have strong experience in SAS programming, PySpark, and cloud platforms like AWS, with a focus on performance optimization and data engineering. This critical role supports advanced analytics capabilities and modernizes data infrastructure.

Qualifications

  • 3+ years of experience in SAS programming and data engineering.
  • 2+ years of hands-on experience with PySpark and big data technologies.
  • Strong understanding of data structures, algorithms, and performance tuning.

Responsibilities

  • Analyze and reverse engineer existing SAS programs.
  • Design and implement equivalent PySpark modules optimized for performance.
  • Develop and maintain ETL pipelines using PySpark.

Skills

SAS programming
PySpark
Data engineering
Cloud platforms
Performance tuning

Education

Bachelor’s degree in Computer Science or related field

Tools

AWS
Snowflake
DataBricks
Job description
Overview

Role Overview: We are seeking a skilled and motivated SAS to PySpark Developer to lead the migration of legacy SAS-based analytics workflows to scalable PySpark solutions. This role is critical to modernizing our data infrastructure and enabling advanced analytics capabilities across cloud platforms.

Responsibilities
  • Analyze and reverse engineer existing SAS programs, macros, and data flows.
  • Design and implement equivalent PySpark modules optimized for performance and scalability.
  • Develop and maintain ETL pipelines using PySpark integrated with AWS, Snowflake/DataBricks
  • Validate migrated outputs against SAS baselines to ensure accuracy and consistency.
  • Collaborate with cross-functional teams including data scientists, business analysts, and cloud engineers.
  • Document technical specifications, migration strategies, and validation protocols.
  • Participate in sprint planning, estimation, and delivery of migration milestones.
Required Qualifications
  • Bachelor’s degree in Computer Science, Engineering, or related field.
  • 3+ years of experience in SAS programming and data engineering.
  • 2+ years of hands-on experience with PySpark and big data technologies.
  • Strong understanding of data structures, algorithms, and performance tuning.
  • Experience with cloud platforms (AWS, Azure, Snowflake) and distributed computing.
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.