1 day ago Be among the first 25 applicants
Direct message the job poster from CyberSolve IT Inc.
Co-Founder | President, CyberSolve IT Inc.
Position Title : Senior Data Architect (GCP – Lakehouse, AI / ML)
Employment Type : Contract to hire (Client Looking to hire only Visa Independent Candidates GC, Citizen etc)
Start Date : Immediate
About the Role
We are seeking an experienced and highly skilled Data Architect to join our dynamic team and lead the development of next-generation cloud-based data platforms. This role is ideal for a strategic, hands-on technical leader with deep expertise in Google Cloud Platform (GCP), Lakehouse architectures, and data engineering. You will help shape the future of data strategy in a leading healthcare organization focused on data-driven decision-making, operational efficiency, and better patient outcomes.
Key Responsibilities
Architecture & Technical Leadership
- Design and implement scalable, high-performance, cost-effective data architecture solutions using GCP technologies: BigQuery, Dataflow, Dataproc, Cloud Spanner, Pub/Sub, GCS, Vertex AI.
- Architect and manage data lakes/warehouses, with strong emphasis on Lakehouse principles and technologies: Delta Lake, Apache Iceberg, Hudi.
- Lead the development of data ingestion, transformation (ETL / ELT) pipelines across structured and unstructured data sources.
Governance, Standards, and Strategy
- Define and enforce data architecture best practices, including data governance, security, retention, and compliance.
- Develop documentation and artifacts to illustrate the data lifecycle, from ingestion through consumption.
- Provide thought leadership and contribute to enterprise-wide data strategy initiatives.
- Guide and mentor data engineers and junior architects.
Collaboration & Stakeholder Engagement
- Work with business stakeholders to translate strategic goals into practical data solutions.
- Collaborate cross-functionally with software engineers, DevOps, product teams, and analysts to ensure data systems meet end-user needs.
- Maintain strong communication with data governance, compliance, and security teams.
Required Skills & Experience
- 8+ years of experience in data architecture, engineering, and data management.
- 5+ years of GCP experience, including BigQuery, Cloud Storage, Pub/Sub, Dataflow, Dataproc, Cloud Composer.
- Proven experience designing Lakehouse architectures using Delta Lake, Iceberg, or Hudi.
- Strong knowledge of schema evolution, data partitioning, indexing, ACID compliance, and distributed file systems.
- Proficient in Python, SQL, and familiarity with Apache Spark, Airflow, and CI/CD pipelines.
- Deep understanding of MLOps, real-time data processing, and integrating AI/ML into data workflows.
- Strong analytical and problem-solving skills with a business mindset.
- Familiar with BI/AI tools and their integration with modern data platforms (e.g., Looker, Power BI, Tableau, Vertex AI).
- Hands-on experience with data modeling, metadata management, and data quality frameworks.
- Experience in Agile/Scrum environments.
Preferred Qualifications
- Experience in healthcare or regulated data environments.
- Exposure to FHIR, HL7, or other healthcare data standards.
- Experience with Apache Beam, Kafka, or other streaming platforms.
- Familiarity with React, Dash, or front-end tools for visualizing data pipelines (a plus).
Core Competencies
- Excellent communication and interpersonal skills.
- Strategic thinking and technology foresight.
- Strong project management and multitasking capabilities.
- Ability to work independently and drive outcomes across teams.
Seniority level: Mid-Senior level
Employment type: Contract
Job function: Consulting
Industries: IT Services and IT Consulting
Note: The job posting appears active; no indication of expiration was found.