¡Activa las notificaciones laborales por email!

Ingeniero de datos

The Flock

México

A distancia

USD 80,000 - 120,000

Jornada completa

Ayer
Sé de los primeros/as/es en solicitar esta vacante

Descripción de la vacante

A leading healthcare technology company is hiring a Sr. Big Data Engineer for a 100% remote role based in Mexico. The role involves ensuring data quality across large-scale systems, collaborating with international teams, and designing automated data quality pipelines primarily on Google Cloud Platform. Ideal candidates will have strong Python, SQL, and GCP experience, along with excellent communication skills.

Formación

  • Strong experience with Python and/or Scala.
  • Solid command of SQL, Hive (HQL), Spark/PySpark.
  • Proven hands-on experience with data quality checks and validation frameworks.
  • Comfortable with structured and unstructured data formats like JSON and Parquet.
  • Excellent communication and stakeholder management skills.

Responsabilidades

  • Ensure data quality, reliability, and integrity in large-scale systems.
  • Design, code, deploy, and monitor automated data quality pipelines.
  • Collaborate with international teams on data quality processes.

Conocimientos

Python
SQL
GCP environments
Data quality checks
Stakeholder management
Shell scripting
Tableau

Herramientas

Google Cloud Platform
GitHub
Terraform

Descripción del empleo

We are hiring a Sr. Big Data Engineer – Data Quality

100% remote – must be based in Mexico

Industry: Healthcare

Full-time position under payroll - Long term

About the role

Join the Data & AI team, where you'll play a key hands-on role in ensuring the quality, reliability, and integrity of data across large-scale systems. You’ll work in a healthcare environment with a Salesforce-centered ecosystem, collaborating with Data Engineers, Product Managers, and business teams to implement robust data quality processes.

You will interact with a distributed international team, including engineers based in India, so strong communication and stakeholder management skills are essential.

This is a highly technical role: we are looking for someone who can design, code, deploy, and monitor automated data quality pipelines, primarily on Google Cloud Platform (GCP).

Must-have skills

  • Strong experience with Python and/or Scala
  • Solid command of SQL, Hive (HQL), Spark/PySpark
  • Proven hands-on experience with data quality checks, validation frameworks, and automation
  • Experience working in GCP environments (priority over other clouds)
  • Comfortable with structured and unstructured data (e.g., JSON, Parquet, ORC, AVRO)
  • Fluent in Linux-based systems
  • Familiarity with CI/CD practices and GitHub
  • Experience building dashboards using Tableau (or similar tools)
  • Excellent communication and stakeholder management skills
  • Detail-oriented and self-driven mindset

Nice to have

  • Experience with Terraform or other Infrastructure-as-Code tools
  • Familiarity with shell scripting
  • Previous experience in data engineering, QA, or data testing roles
  • Performance tuning and troubleshooting of data pipelines
  • Experience in Salesforce data models is a plus
  • Willingness to obtain certifications in the near future (company-supported)

Who we’re looking for

A highly technical and pragmatic engineer who’s eager to get hands-on with data quality challenges. You should be comfortable diving into complex data workflows, solving problems independently, and communicating clearly with international teams.

If you're ready to bring clean, reliable data to life in a mission-critical healthcare setting — we want to hear from you.

Apply now or refer someone you trust:

https://app.theflock.com/f/Openings/PublicOpening/515

or send me an E-mail: leticia.mastroberti@theflock.com

#BigData #DataQuality #Spark #Python #GCP #SQL #Hiring #HealthcareData #RemoteJobs #MexicoJobs #DataEngineer

Consigue la evaluación confidencial y gratuita de tu currículum.
o arrastra un archivo en formato PDF, DOC, DOCX, ODT o PAGES de hasta 5 MB.