Job Search and Career Advice Platform

Ativa os alertas de emprego por e-mail!

Data Scientist & Engineering Lead

Bebeedata

Teletrabalho

BRL 160.000 - 200.000

Tempo integral

Ontem
Torna-te num dos primeiros candidatos

Cria um currículo personalizado em poucos minutos

Consegue uma entrevista e ganha mais. Sabe mais

Resumo da oferta

A leading data solutions company is seeking a highly skilled Data Scientist and Engineering Lead. This is an exceptional opportunity for those with experience in building scalable data pipelines using Spark, PySpark, SQL, and Delta Lake. The role includes designing ETL pipelines, crafting efficient data ingestion processes, and delivering insights through Power BI dashboards. It offers the flexibility of a remote contract while being part of a dynamic team.

Qualificações

  • Strong background in building scalable data pipelines using Spark, PySpark, SQL, and Delta Lake.
  • Hands-on experience in ingesting data from multiple sources and transforming it for analytics.
  • Ability to deliver high-quality datasets and visualizations.

Responsabilidades

  • Design and implement robust ETL / ELT pipelines using Databricks, ensuring optimal performance.
  • Craft efficient data ingestion processes from diverse sources.
  • Develop reusable pipeline frameworks and data validation logic.
  • Create curated datasets and deliver actionable insights through dashboards.
  • Analyze pipeline performance issues and drive process improvements.

Conhecimentos

Spark
PySpark
SQL
Delta Lake
Power BI
Descrição da oferta de emprego

Data Scientist & Engineering Lead

We are seeking a highly skilled Data Scientist and Engineering Lead to spearhead our data engineering efforts. This is an exceptional opportunity for someone with a strong background in building scalable data pipelines using Spark, PySpark, SQL, and Delta Lake.

The ideal candidate will have hands‑on experience ingesting data from multiple sources, transforming it for analytics, and publishing high‑quality datasets and visualizations.

Responsibilities

Design and implement robust ETL / ELT pipelines in Databricks using PySpark, Spark SQL, and Delta Lake, ensuring seamless data flow and optimal performance.

Craft efficient data ingestion processes from diverse sources (APIs, SQL databases, cloud storage, SAP / legacy systems, streaming), guaranteeing accurate and timely data delivery.

Develop reusable pipeline frameworks, data validation logic, and performance‑tuned transformations, fostering a culture of collaboration and innovation.

Create curated datasets and deliver actionable insights through Power BI dashboards, empowering business stakeholders with data‑driven decisions.

Establish best practices for lakehouse development, orchestration, and version control, ensuring scalability, reliability, and maintainability.

Analyze pipeline performance issues, resolve data accuracy and quality concerns, and drive process improvements, upholding the highest standards of data excellence.

About This Role

This is a remote contract position that offers the flexibility to work independently while being part of a dynamic team.

Obtém a tua avaliação gratuita e confidencial do currículo.
ou arrasta um ficheiro em formato PDF, DOC, DOCX, ODT ou PAGES até 5 MB.