Overview
We are seeking experienced Data Engineers for multiple roles across healthcare and enterprise data platforms. The positions involve modernizing on-prem data warehouses to cloud environments, building data marts on Snowflake/AWS, and delivering reliable ETL/ELT pipelines, data governance, and reporting solutions. Roles vary from contractor to full-time opportunities with long-term project pipelines. This description consolidates responsibilities, qualifications, and benefits from the provided content while removing irrelevant boilerplate.
Responsibilities
- Design, build, and maintain data pipelines and ETL/ELT processes to ingest and transform data from diverse sources.
- Develop and optimize SQL, stored procedures, views, and data models to support enterprise reporting and analytics.
- Collaborate with stakeholders to document data sources, transformations, and dependencies; provide data governance and quality checks.
- Support BI developers with clean, structured datasets and operational dashboards.
- Implement automated QA checks, monitoring, and root-cause analysis for data pipelines; maintain documentation of operational procedures.
- On AWS and/or Azure environments, work with services like S3, Glue, SageMaker, Data Factory, Synapse, BigQuery, DBT, and Looker/Power BI integration as applicable.
- Contribute to continuous improvement: automate manual tasks, optimize pipelines for performance and cost, and implement scalable architectures.
- Provide knowledge transfer and mentorship to internal teams; participate in code reviews and cross-team collaboration.
Required Qualifications
- 5+ years (or as specified per role) of data engineering experience with cloud data platforms (AWS, Azure, GCP) and building scalable data pipelines.
- Strong proficiency in Python and SQL; deep experience with data modeling (dimensional modeling, star schemas, data warehouses).
- Experience with ETL/ELT tooling (e.g., Airflow, dbt, AWS Glue, Azure Data Factory, Synapse).
- Experience with data governance, data quality frameworks, and metadata management.
- Familiarity with integrating data from databases, APIs, and external data providers; ability to design normalized schemas and choose between SQL/NoSQL as needed.
- Excellent problem-solving, communication, and collaboration skills; ability to work in Agile environments and across teams.
Nice-to-Have
- Experience with healthcare IT data, BigQuery, Looker, Tableau, Power BI, or similar BI tools.
- Hands-on experience with containerization (Docker, Kubernetes) and orchestration; familiarity with Terraform or IaC.
- Experience with ML/AI workflows (e.g., SageMaker, BQML) and data validation frameworks (e.g., Great Expectations).
- Experience with data streaming (Kafka, Spark) and real-time processing; knowledge of data lakehouse concepts.
What to Expect
- Contract durations vary (e.g., 4–6 months with potential extension) and may involve long-term project work.
- Remote-friendly arrangements with varying time zones; opportunities to work with global teams.
What We Offer
- Competitive compensation; remote or hybrid opportunities depending on role.
- Benefits and professional development programs as applicable to contract roles and full-time positions.
- Access to learning portals, certifications, and career growth paths.
We are seeking a Data Engineer to design and maintain scalable data pipelines on AWS/Azure, ensuring performance, quality, and security. The engineer will collaborate with data scientists and analysts to integrate data from multiple sources and support analytics initiatives. This description preserves the essential content while improving readability and compliance with formatting guidelines.
Additional notes: The original content contains varied postings and client contexts. This refined description focuses on core responsibilities, qualifications, and benefits applicable to a data engineering role, with emphasis on cloud platforms, data pipelines, governance, and collaboration. It retains references to healthcare and enterprise data environments where relevant while omitting extraneous promotional material.