Job Description
Position: Senior Data Engineer
Resume Due Date: Wednesday, June 25, 2025 (5:00PM EST)
Number of Vacancies: 2
Level: MP4 upto $90/hr INC
Duration: 12 Months
Hours of work: 35
Location: 889 Brock Road, Pickering (Hybrid – 4 days remote)
Job Overview
- As a Senior Data Developer, you will be responsible for building and supporting data-driven applications that enable innovative, customer-centric digital experiences.
- You will work as part of a cross-disciplinary agile team to solve problems across all business areas.
- You will build reliable, supportable, and performant data lake and data warehouse products to meet organizational needs for data reporting, analytics, applications, and innovation.
- You will employ best practices in development, security, and accessibility to ensure high-quality service for our customers.
- Build and productionize modular and scalable data ELT/ETL pipelines and infrastructure, leveraging diverse data sources across the organization.
- Implement data ingestion and curation pipelines that provide a unified, business-centric source of truth for BI, reporting, and downstream systems, in collaboration with Data Architect.
- Work closely with Data Architect, infrastructure, and cybersecurity teams to ensure data security in transit and at rest.
- Clean, prepare, and optimize datasets for performance, ensuring lineage and quality controls throughout the data integration cycle.
- Support Business Intelligence Analysts in data modeling for visualization and reporting using dimensional data modeling and aggregation techniques.
- Provide production support for ingestion, transformation, pipeline performance, data accuracy, and integrity issues.
- Collaborate with data architect, business analysts, data scientists, data engineers, data analysts, solution architects, and data modelers to develop data pipelines for our data marketplace.
- Assist in identifying, designing, and implementing process improvements such as automation, optimization, and infrastructure re-design for scalability.
- Work with tools in the Microsoft Stack including Azure Data Factory, Azure Data Lake, Azure SQL Databases, Azure Data Warehouse, Azure Synapse Analytics, Azure Databricks, Collibra, and Power BI.
- Participate in agile SCRUM processes, contributing to backlog items and utilizing Kanban/SCRUM tools.
- Assist in building and maintaining a data catalog and relevant metadata for enterprise datasets.
- Develop optimized, scalable data pipelines and models using technologies like Python, Spark, and SQL, consuming data in formats such as XML, CSV, JSON, REST APIs.
- Document data pipelines and products thoroughly, using source control for maintainability.
- Implement orchestration to ensure data products meet latency and dependency requirements with minimal disruption.
- Create tools to automate routine tasks and reduce manual effort.
- Integrate with CI/CD and DevOps pipelines for automation, release management, and version control.
- Monitor in-production solutions, troubleshoot issues, and provide Tier 2 support as needed.
- Manage data access via role-based controls.
- Write and execute automated testing, assist in user acceptance and system testing, and help design test cases.
- Participate in peer code reviews and approve pull requests in non-production environments.
Qualifications
- Four-year university degree in computer science, software engineering, or related field, with focus on data engineering, analysis, AI, or ML.
- Experience designing and building data pipelines using Azure Data Factory and Databricks.
- Proficiency in Python, PySpark, SparkSQL, and SQL for data processing frameworks.
- Experience with Azure Data Factory, ADLS, Synapse Analytics, and Databricks.
- Experience developing data pipelines for Data Lakehouses and Data Warehouses.
- Strong understanding of data structures and processing frameworks.
- Knowledge of data governance and data quality principles.
- Effective communication skills to translate technical concepts to non-technical stakeholders.