Ativa os alertas de emprego por e-mail!

Data Engineer (Collibra Data Quality)

buscojobs Brasil

Pará

Teletrabalho

BRL 120.000 - 160.000

Tempo integral

Hoje
Torna-te num dos primeiros candidatos

Resumo da oferta

A leading job marketplace is seeking a Data Engineer to create and maintain scalable data pipelines using AWS or Azure. The ideal candidate will have over 5 years of experience and expertise in Python, SQL, and ETL/ELT tools. This role offers competitive compensation and opportunities for remote or hybrid work arrangements.

Serviços

Competitive compensation
Remote or hybrid work options
Access to learning portals and certifications

Qualificações

  • 5+ years of data engineering experience with cloud data platforms.
  • Strong proficiency in Python and SQL with experience in data modeling.
  • Experience in ELT/ETL tooling and data governance frameworks.

Responsabilidades

  • Design and maintain data pipelines and ETL/ELT processes.
  • Collaborate with stakeholders and provide data governance.
  • Implement automated QA checks for data pipelines.

Conhecimentos

Python
SQL
Data Modeling
ETL/ELT tooling
Data Governance
Problem-solving
Communication
Collaboration

Ferramentas

AWS
Azure
GCP
Airflow
Docker
Kubernetes
Terraform
Descrição da oferta de emprego
Overview

We are seeking experienced Data Engineers for multiple roles across healthcare and enterprise data platforms. The positions involve modernizing on-prem data warehouses to cloud environments, building data marts on Snowflake/AWS, and delivering reliable ETL/ELT pipelines, data governance, and reporting solutions. Roles vary from contractor to full-time opportunities with long-term project pipelines. This description consolidates responsibilities, qualifications, and benefits from the provided content while removing irrelevant boilerplate.

Responsibilities
  • Design, build, and maintain data pipelines and ETL/ELT processes to ingest and transform data from diverse sources.
  • Develop and optimize SQL, stored procedures, views, and data models to support enterprise reporting and analytics.
  • Collaborate with stakeholders to document data sources, transformations, and dependencies; provide data governance and quality checks.
  • Support BI developers with clean, structured datasets and operational dashboards.
  • Implement automated QA checks, monitoring, and root-cause analysis for data pipelines; maintain documentation of operational procedures.
  • On AWS and/or Azure environments, work with services like S3, Glue, SageMaker, Data Factory, Synapse, BigQuery, DBT, and Looker/Power BI integration as applicable.
  • Contribute to continuous improvement: automate manual tasks, optimize pipelines for performance and cost, and implement scalable architectures.
  • Provide knowledge transfer and mentorship to internal teams; participate in code reviews and cross-team collaboration.
Required Qualifications
  • 5+ years (or as specified per role) of data engineering experience with cloud data platforms (AWS, Azure, GCP) and building scalable data pipelines.
  • Strong proficiency in Python and SQL; deep experience with data modeling (dimensional modeling, star schemas, data warehouses).
  • Experience with ETL/ELT tooling (e.g., Airflow, dbt, AWS Glue, Azure Data Factory, Synapse).
  • Experience with data governance, data quality frameworks, and metadata management.
  • Familiarity with integrating data from databases, APIs, and external data providers; ability to design normalized schemas and choose between SQL/NoSQL as needed.
  • Excellent problem-solving, communication, and collaboration skills; ability to work in Agile environments and across teams.
Nice-to-Have
  • Experience with healthcare IT data, BigQuery, Looker, Tableau, Power BI, or similar BI tools.
  • Hands-on experience with containerization (Docker, Kubernetes) and orchestration; familiarity with Terraform or IaC.
  • Experience with ML/AI workflows (e.g., SageMaker, BQML) and data validation frameworks (e.g., Great Expectations).
  • Experience with data streaming (Kafka, Spark) and real-time processing; knowledge of data lakehouse concepts.
What to Expect
  • Contract durations vary (e.g., 4–6 months with potential extension) and may involve long-term project work.
  • Remote-friendly arrangements with varying time zones; opportunities to work with global teams.
What We Offer
  • Competitive compensation; remote or hybrid opportunities depending on role.
  • Benefits and professional development programs as applicable to contract roles and full-time positions.
  • Access to learning portals, certifications, and career growth paths.

We are seeking a Data Engineer to design and maintain scalable data pipelines on AWS/Azure, ensuring performance, quality, and security. The engineer will collaborate with data scientists and analysts to integrate data from multiple sources and support analytics initiatives. This description preserves the essential content while improving readability and compliance with formatting guidelines.

Additional notes: The original content contains varied postings and client contexts. This refined description focuses on core responsibilities, qualifications, and benefits applicable to a data engineering role, with emphasis on cloud platforms, data pipelines, governance, and collaboration. It retains references to healthcare and enterprise data environments where relevant while omitting extraneous promotional material.

Obtém a tua avaliação gratuita e confidencial do currículo.
ou arrasta um ficheiro em formato PDF, DOC, DOCX, ODT ou PAGES até 5 MB.