
Enable job alerts via email!
Generate a tailored resume in minutes
Land an interview and earn more. Learn more
A leading technology services firm is seeking a Cloud Data Architect to define and own enterprise data architecture patterns, including lakehouse, data lake, and data governance. The successful candidate will create reference architectures for data pipelines using Databricks and AWS services, focusing on performance optimization and reliability standards. This role requires expertise in governance, security, and orchestration tools like Airflow. Excellent opportunity to lead transformative data initiatives in a collaborative environment.
Define and own enterprise data architecture patterns: lakehouse, data warehouse, data lake, data vault and dimensional models, aligned to business needs and regulatory requirements.
Create reference architectures for batch, streaming, and transactional pipelines using Databricks (DLT, Autoloader, Unity Catalog, SQL Warehouse).
Establish performance optimization guidelines for PySpark/Spark: memory tuning, shuffle/partition strategies, UDF optimization, RAPIDS/GPU acceleration.
Design event‑driven ingestion and CDC architectures (GoldenGate, Kafka/MSK, Kinesis, Glue/Airflow operators).
Implement data governance: Unity Catalog, access controls, lineage, PII handling, encryption/decryption in transit and at rest.
Define observability & reliability standards: data quality (DQ), schema evolution, incident management, SLAs/SLOs, cost guardrails.
Standardize Airflow/MWAA/Composer orchestration with reusable operators and DAG patterns.