Enable job alerts via email!

Data Architect

Cpus Engineering Staffing Solutions Inc.

Toronto

Remote

CAD 90,000 - 130,000

Full time

3 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

A leading engineering staffing solutions company seeks a Data Architect to design and lead data infrastructure implementations. This remote position entails developing scalable data solutions, guiding data pipeline strategies, and collaborating with various teams to enhance data quality and accessibility. Ideal candidates will have a strong educational background and extensive experience in data modeling, cloud architectures, and event-driven data solutions.

Qualifications

  • Minimum of 6-8 years experience in data modeling and data architecture.
  • Experience in cloud environments and various data formats.
  • Ability to design event-driven data solutions.

Responsibilities

  • Lead the design and implementation of data pipelines.
  • Collaborate with cross-functional teams to enhance data delivery.
  • Ensure data security and optimal performance.

Skills

Data modeling
Data architecture
Data pipeline design
Cloud solutions
Big Data knowledge
Event-driven data solutions

Education

4 years University education in computer science

Tools

Azure Data Factory
Azure Data Lake
Azure SQL Databases
Azure Data Warehouse
Azure Synapse Analytics Services
Azure Databricks
Power BI

Job description

We are currently requesting resumes for the following position : Data Architect

Number of Vacancies : 2

Job ID : 24-079

Level : MP6

Duration : 4 months

Hours of work : 40

Location : 700 University Ave (100% Remote)

Job Overview

  • Lead the architecture, design, and oversee implementation of modular and scalable data ELT / ETL pipelines and data infrastructure leveraging the wide range of data sources across the organization.
  • Design curated common data models that offer an integrated, business-centric single source of truth for business intelligence, reporting, and downstream system use.
  • Work closely with infrastructure and cyber teams to ensure data is secure in transit and at rest.
  • Create, guide, and enforce code templates for delivery of data pipelines and transformations for structured, semi-structured and unstructured data sets.
  • Develop modeling guidelines that ensure model extensibility and reuse by employing industry standard disciplines for building facts, dimensions, bridge, aggregates, slowly changing dimensions, and other dimensional and fact optimizations.
  • Establish standards database system fields, including primary and natural key combinations that optimize join performance in a multi-domain, multiple subject area physical (structured zone) and semantic model (curated zone)
  • Ensure model extensibility by employing industry standard disciplines for building facts, dimensions, bridge, aggregates, slowly changing dimensions, and other dimensional and fact optimizations.
  • Transform data and map to more valuable and understandable semantic layer sets for consumption, transitioning from system centric language to business-centric language.
  • Collaborate with business analysts, data scientists, data engineers, data analysts and solution architects to develop data pipelines to feed our data marketplace.
  • Introduce new technologies to the environment through research and POCs and prepare POC code designs that can be implemented and productionized by developers.
  • Work with tools in the Microsoft Stack; Azure Data Factory, Azure Data Lake, Azure SQL Databases, Azure Data Warehouse, Azure Synapse Analytics Services, Azure Databricks, Microsoft Purview, and Power Bl
  • Work within the agile SCRUM work management framework in delivery of products and services, including contributing to feature & user story backlog item development, and utilizing related Kanban / SCRUM toolsets.
  • Document as-built architecture and designs within the product description.
  • Design data solutions that enable batch, near-real-time, event-driven, and / or streaming approaches depending on business requirements
  • Design & advise on orchestration of data pipeline execution to ensure data products meet customer latency expectations, dependencies are managed, and datasets are as up to date as possible, with minimal disruption to end-customer use.
  • Ensure that designs are implemented with proper attention to data security, access management, and data cataloging requirements.
  • Approve pull requests related to production deployments.
  • Demonstrate solutions to business customers to ensure customer acceptance and solicit feedback to drive iterative improvements.
  • Assist in troubleshooting issues for datasets produced by the team (Tier 3 support), on an as required basis.
  • Guide data modelers, business analysts and data scientists in the build of models optimized for KPI delivery, actionable feedback / writeback to operational systems and enhancing the predictability of machine learning models and experiments.

Qualifications

  • 4 years year University education in computer science, software engineering or other relevant programs within data engineering, data analysis, artificial intelligence, or machine learning.
  • Minimum experience of +6 years and up to and including 8 years of experience is required in data modeling, data warehouse design, and data solution architecture in a Big Data environment is considered necessary.
  • Experience guiding data lake ingestion and data modeling projects in a cloud environment, experience in modeling relational and in-memory models with star / snowflake schemas.
  • Experience with designing and implementing event-driven (pub / sub), near-real-time, or streaming data solutions, involving structured, semi-structured and unstructured data across various platforms.
  • Requires an extensive knowledge in designing a data model to solve a business problem, specifying a data pipeline design pattern to bring data into a data warehouse, optimizing data structures to achieve required performance, designing low-latency and / or event-driven patterns of data processing, creation of a common data model to support current and future business needs.

J-18808-Ljbffr

Create a job alert for this search
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Principal Data Architect

Carbon60

Toronto

Remote

CAD 120,000 - 180,000

2 days ago
Be an early applicant

Databricks Data Architect, Deloitte Global Technology

Deloitte Canada

Toronto

Remote

CAD 85,000 - 156,000

19 days ago

Senior Data Solutions Architect

Opsguru

Toronto

Remote

CAD 100,000 - 130,000

2 days ago
Be an early applicant

Principal Data Architect

OpsGuru

North Vancouver

Remote

CAD 120,000 - 150,000

Yesterday
Be an early applicant

Data Architect

Bridgenext

Ottawa

Remote

CAD 90,000 - 130,000

Yesterday
Be an early applicant

Principal Data Architect

Lumenalta (formerly Clevertech)

Ottawa

Remote

CAD 90,000 - 130,000

2 days ago
Be an early applicant

Principal Software Architect

OpsGuru

Ontario

Remote

CAD 110,000 - 160,000

2 days ago
Be an early applicant

Senior Data Solutions Architect

Carbon60

Toronto

Remote

CAD 100,000 - 130,000

4 days ago
Be an early applicant

Data Architect

Mpa Recruitment

Halifax

Remote

CAD 90,000 - 130,000

4 days ago
Be an early applicant