Enable job alerts via email!

Data Platform Engineer

Vireo Health

United States

Remote

USD 90,000 - 150,000

Full time

6 days ago
Be an early applicant

Boost your interview chances

Create a job specific, tailored resume for higher success rate.

Job summary

Join a forward-thinking company on a mission to transform healthcare through data-driven solutions. We are seeking skilled data engineers to build a robust data infrastructure that supports ambitious growth and innovation in AI and analytics. In this role, you will design scalable data solutions, optimize data management processes, and collaborate with cross-functional teams to enhance our platform capabilities. With a commitment to talent development, this position offers exciting opportunities for growth and the potential to transfer to our US office. If you're passionate about leveraging data to drive impactful decisions, this is the perfect opportunity for you.

Benefits

Great compensation
Stock options in a PUBLIC company (ESOP)
13th month salary
Opportunity to transfer to the US
Work with LLMs

Qualifications

  • 4+ years in data platform engineering or backend development.
  • Experience with SQL for data transformation and analysis.
  • Strong programming skills in Python, Go, or Java.

Responsibilities

  • Support building scalable data infrastructure for acquisitions.
  • Design and optimize data warehouse and lakehouse solutions.
  • Develop ETL/ELT pipelines for data management and governance.

Skills

Data Platform Engineering
SQL
Python
Go
Java
Distributed Data Processing
ETL Workflows
Data Integration
Problem Solving
Communication Skills

Education

Bachelor's Degree in Computer Science or related field

Tools

AWS
GCP
Azure
Apache Spark
Apache Flink
Docker
Kubernetes
DBT
Airflow
Kafka

Job description

With annual revenues of $400M, Vireo Health is transforming the healthcare landscape through our ambitious mission to focus on data to drive our business and to build a data infrastructure that helps us scale through acquisitions, aiming to grow from $400M to $1B+ in 2 years. We are building a highly sophisticated business operation utilizing data for LLMs, ML, and analytics. Our Saigon AI and data platform team is expanding, and we are looking to add two data engineers who can help us develop a well-governed data infrastructure from an early stage system and architecture. We are also leveraging LLMs to inform strategic and operational decisions, seeking candidates committed to learning and growth in this area.

Commitment to Talent: Excellent candidates may transfer to our US office after working with us for >1 year (if they wish).

Responsibilities

  • Support our solutions architect and data platform leader in building a scalable data infrastructure to support acquisitions of 4-5 companies per year, deployment of AI/ML, and scaling delivery and retail sales from $400M to $1.5B.

Data Warehouse & Lakehouse:

  • Design, implement, and optimize scalable data warehouse and lakehouse solutions (e.g., Delta Lake, Apache Iceberg, Snowflake, BigQuery).
  • Develop ETL/ELT pipelines for efficient ingestion, transformation, and management of structured and unstructured data.
  • Ensure data governance, lineage, and security best practices, including access controls and encryption.
  • Optimize data storage, partitioning, and query performance for cost efficiency.

Data Platform, Cloud & System Reliability:

  • Build and maintain a robust, scalable data platform supporting engineering, analytics, and AI workloads.
  • Deploy, manage, and optimize data infrastructure across cloud platforms (AWS, GCP, Azure).
  • Enhance system reliability, observability, and cost efficiency through monitoring, logging, and automation.
  • Develop SDKs, APIs, and automation tools to improve data engineering workflows.
  • Collaborate with Data Analytics, ML Engineers, and Software Engineers to enhance platform capabilities and scalability.
  • Integrate data with AI workloads, leveraging LLM techniques such as retrieval-augmented generation (RAG) and fine-tuning.
  • Index and manage data within Vector Stores for real-time AI applications.
  • Develop external tools and APIs to enable LLMs and AI models to query and interact with data efficiently.
Your skills and experience
  • 4+ years of experience in data platform engineering, backend development, or data infrastructure.
  • Ability to craft complex SQL queries for data transformation, analysis, and optimization.
  • Strong programming skills in Python, Go, or Java for building scalable data solutions and microservices.
  • Deep knowledge of distributed data processing frameworks (e.g., Spark, Flink) and hands-on experience with at least one cloud data platform (AWS Glue + Athena, Google BigQuery, Snowflake, or Databricks).
  • Proven experience in building reliable data pipelines and ETL workflows using workflow orchestration tools (e.g., DBT, Airflow, Dagster, Prefect).
  • Strong experience in data integration across various sources, including RDBMS, Vector Stores, and real-time data streaming technologies (Kafka, Pulsar, Kinesis).
  • Proficiency in containerization with Kubernetes, Docker for scalable Data & AI workloads.
  • Experience designing and optimizing Data Warehouse and Data Lakehouse Architectures: Delta Lake, Apache Iceberg.
  • Strong problem-solving skills, a product-led mindset, and AI-first thinking, with experience using AI for coding.
  • Excellent communication skills in English, with the ability to collaborate across technical and non-technical teams.
  • Ability to quickly prototype, test, and iterate solutions and products.

Preferred Qualifications

  • Familiarity with vector databases (Pinecone, FAISS, Weaviate, ChromaDB) for AI-driven applications is a plus.
  • Understanding of data versioning, lineage, and governance (e.g., Apache Atlas, Great Expectations).

Why you'll love working here

  • Great compensation
  • Stock options in a PUBLIC company (ESOP)
  • 13th month and more
  • Opportunity to transfer to the US
  • Work with LLMs
Get your free, confidential resume review.
or drag and drop a PDF, DOC, DOCX, ODT, or PAGES file up to 5MB.

Similar jobs

Senior Data Platform Engineer (Remote)

AVAILITY

Remote

USD 90,000 - 140,000

4 days ago
Be an early applicant

Senior Data Platform Engineer (Remote)

Availity, LLC

Remote

USD 90,000 - 150,000

5 days ago
Be an early applicant

Data Platforms Engineer

Blank Rome LLP

Remote

USD 100,000 - 115,000

27 days ago

Data Platform Engineer

Motive

Indianapolis

Remote

USD 90,000 - 130,000

2 days ago
Be an early applicant

Senior Data Platform Engineer (Remote)

Availity

Remote

USD 110,000 - 150,000

3 days ago
Be an early applicant

Data Platform Engineer

Upstart

Remote

USD 142,000 - 197,000

30+ days ago

Data Platform Engineer

Upstart

Remote

USD 142,000 - 197,000

30+ days ago

Data Platform Engineer

Ohiox

Columbus

Remote

USD 142,000 - 197,000

30+ days ago

[Hiring] Data Platform Engineer @Twonice

Twonice

Remote

CAD 130,000 - 165,000

9 days ago