Ativa os alertas de emprego por e-mail!

Sr. Data Architect

buscojobs Brasil

Caraí

Teletrabalho

BRL 160.000 - 200.000

Tempo integral

Ontem
Torna-te num dos primeiros candidatos

Resumo da oferta

A technology-focused solutions provider is seeking a Senior Data Engineer to design and maintain scalable data pipelines on AWS. The role involves collaborating with data scientists, optimizing ETL processes, and ensuring high data quality. Applicants should have over 5 years of experience, excellent problem-solving skills, and strong programming knowledge in Python and SQL. This position offers a fully remote work environment and competitive pay.

Serviços

100% Remote Work
WFH allowance
Career Growth opportunities

Qualificações

  • 5+ years of experience building scalable and reliable data pipelines in a cloud environment.
  • Deep understanding of ELT processes and data modeling best practices.
  • Excellent problem-solving abilities and strong attention to data accuracy and detail.

Responsabilidades

  • Designing performant data pipelines for complex datasets.
  • Building scalable infrastructure for data updates.
  • Implementing automated QA checks and monitoring systems.

Conhecimentos

Data engineering with AWS
Python
SQL
ETL processes
Data modeling
Advanced English

Formação académica

Bachelor's degree in Computer Science or related field

Ferramentas

AWS Glue
Docker
BigQuery
Descrição da oferta de emprego
Overview

We are looking for a Senior Data Engineer to design and maintain scalable data pipelines on AWS, ensuring performance, quality, and security. You will collaborate with data scientists and analysts to integrate data from multiple sources and support AI/ML initiatives.

Responsibilities
  • Build and optimize ETL pipelines with AWS Glue.
  • Work with AWS S3, Glue, and SageMaker for data and AI workflows.
  • Develop solutions in Python and SQL.
  • Integrate data from Salesforce and APIs.
  • Ensure data governance, documentation, and best practices.
  • AWS (S3, Glue, SageMaker)
  • Python, SQL
  • Designing performant data pipelines for the ingestion and transformation of complex datasets into usable data products.
  • Building scalable infrastructure to support hourly, daily, and weekly update cycles.
  • Implementing automated QA checks and monitoring systems to catch data anomalies before they reach clients.
  • Re-architecting system components to improve performance or reduce costs.
  • Supporting team members through code reviews and collaborative development.
  • Building enterprise-grade batch and real-time data processing pipelines on AWS, with a focus on serverless architectures.
  • Designing and implementing automated ELT processes to integrate disparate datasets.
  • Collaborating across multiple teams to ingest, extract, and process data using Python, R, Zsh, SQL, REST, and GraphQL APIs.
  • Transforming clickstream and CRM data into meaningful metrics and segments for visualization.
  • Creating automated acceptance, QA, and reliability checks to ensure business logic and data integrity.
  • Designing appropriately normalized schemas and making informed decisions between SQL and NoSQL solutions.
  • Optimizing infrastructure and schema design for performance, scalability, and cost efficiency.
  • Defining and maintaining CI/CD and deployment pipelines for data infrastructure.
  • Containerizing and deploying solutions using Docker and AWS ECS.
  • Proactively identifying and resolving data discrepancies, and implementing safeguards to prevent recurrence.
  • Contributing to documentation, onboarding materials, and cross-team enablement efforts.
Requirements
  • Proven experience in data engineering with AWS.
  • Experience with ETL, data modeling, and pipeline optimization.
  • Advanced English (international collaboration).
What you’ll bring
  • Bachelor’s degree in Computer Science, Software Engineering, or a related field; additional training in statistics, mathematics, or machine learning is a strong plus.
  • 5+ years of experience building scalable and reliable data pipelines and data products in a cloud environment (AWS preferred).
  • Deep understanding of ELT processes and data modeling best practices.
  • Strong programming skills in Python or a similar scripting language.
  • Advanced SQL skills, with intermediate to advanced experience in relational database design.
  • Familiarity with joining and analyzing large behavioral datasets, such as Adobe and GA4 clickstream data.
  • Excellent problem-solving abilities and strong attention to data accuracy and detail.
  • Proven ability to manage and prioritize multiple initiatives with minimal supervision.
Nice to have
  • Experience working with data transformation tools such as Data Build Tool or similar technologies.
  • Familiarity with Docker containerization and orchestration.
  • Experience in API design or integration for data pipelines.
  • Development experience in a Linux or Mac environment.
  • Exposure to data QA frameworks or observability tools (e.g., Great Expectations, Monte Carlo, etc.).
What we offer
  • 100% Remote Work
  • WFH allowance: Monthly payment as financial support for remote working.
  • Career Growth: We have established a career development program accessible for all employees with a 360º feedback that will help us to guide you in your career progression.
  • Training: For Tech training at Zartis, you have time allocated during the week at your disposal. You can request from a variety of options, such as online courses (from Pluralsight and Educative.io, for example), English classes, books, conferences, and events.
  • Mentoring Program: You can become a mentor in Zartis or you can receive mentorship, or both.
  • Zartis Wellbeing Hub (Kara Connect): A platform that provides sessions with a range of specialists, including mental health professionals, nutritionists, physiotherapists, fitness coaches, and webinars with such professionals as well.
  • Multicultural working environment: We organize tech events, webinars, parties, and activities to do online team-building games and contests.
About The Role

We are seeking experienced Data Engineers to develop and deliver robust, cost-efficient data products that power analytics, reporting and decision-making across two distinct brands.

The project

Our teammates are talented people that come from a variety of backgrounds. We’re committed to building an inclusive culture based on trust and innovation.

You will be part of a distributed team developing new technologies to solve real business problems. Our client empowers organizations to make smarter, faster decisions through the seamless integration of strategy, technology, and analytics. They have helped leading brands harness their marketing, advertising, and customer experience data to unlock insights, enhance performance, and drive digital transformation.

We are looking for someone with good communication skills, ideally with experience making decisions, being proactive, used to building software from scratch, and with good attention to detail.

What you will do
  • Designing performant data pipelines for the ingestion and transformation of complex datasets into usable data products.
  • Building scalable infrastructure to support hourly, daily, and weekly update cycles.
  • Implementing automated QA checks and monitoring systems to catch data anomalies before they reach clients.
  • Re-architecting system components to improve performance or reduce costs.
  • Supporting team members through code reviews and collaborative development.
  • Building enterprise-grade batch and real-time data processing pipelines on AWS, with a focus on serverless architectures.
  • Designing and implementing automated ELT processes to integrate disparate datasets.
  • Collaborating across multiple teams to ingest, extract, and process data using Python, R, Zsh, SQL, REST, and GraphQL APIs.
  • Transforming clickstream and CRM data into meaningful metrics and segments for visualization.
  • Creating automated acceptance, QA, and reliability checks to ensure business logic and data integrity.
  • Designing appropriately normalized schemas and making informed decisions between SQL and NoSQL solutions.
  • Optimizing infrastructure and schema design for performance, scalability, and cost efficiency.
  • Defining and maintaining CI/CD and deployment pipelines for data infrastructure.
  • Containerizing and deploying solutions using Docker and AWS ECS.
  • Proactively identifying and resolving data discrepancies, and implementing safeguards to prevent recurrence.
  • Contributing to documentation, onboarding materials, and cross-team enablement efforts.
What we’re looking for
  • Proven experience building and managing data products in modern cloud environments (GCP preferred).
  • Strong proficiency in Python for data ingestion and workflow development.
  • Hands-on expertise with BigQuery, dbt, Airflow and Looker.
  • Solid understanding of data modeling, pipeline design and data quality best practices.
  • Excellent communication skills and a track record of effective collaboration across technical and non-technical teams.
Why Join Kake?

Kake is a remote-first company with a global community — fully believing that it’s not where your table is, but what you bring to the table. We provide top-tier engineering teams to support some of the world’s most innovative companies, and we’ve built a culture where great people stay, grow, and thrive. We’re proud to be more than just a stop along the way in your career — we’re the destination.

The icing on the Kake

Competitive Pay in USD – Work globally, get paid globally.

Fully Remote – Simply put, we trust you.

Better Me Fund – We invest in your personal growth and passions.

️ Compassion is Badass – Join a community that invests in social good.

Ontem

Cargo

Obtém a tua avaliação gratuita e confidencial do currículo.
ou arrasta um ficheiro em formato PDF, DOC, DOCX, ODT ou PAGES até 5 MB.

Ofertas semelhantes