Location: Toronto, Ontario
Our client is on a mission to power the future of customer support through the responsible and efficient deployment of GenAI bots in contact centers. Our platform transforms raw enterprise data into a secure, structured, and model-ready format, enabling companies to deploy powerful AI assistants with confidence.
They help organizations avoid the costly mistakes of feeding unclean data to LLMs — ensuring their bots don’t hallucinate or mislead customers.
The Role:
We’re looking for a Staff/Data Engineer to join as a key engineering hire. You’ll work closely with the VP of Engineering & AI to build the foundation of our data platform, shaping everything from ingestion to transformation to pipeline orchestration.
This is a hands-on, high-impact role — perfect for someone who thrives in early-stage environments and wants to take full ownership of backend/data systems from the ground up.
What You’ll Do:
- Architect, build, and maintain robust and scalable data pipelines to support GenAI applications.
- Own data ingestion and transformation processes, integrating data from multiple structured and unstructured sources.
- Design and implement Retrieval-Augmented Generation (RAG) systems.
- Work with vector databases, embeddings, and LLM-adjacent infrastructure.
- Collaborate with AI/ML engineers to ensure data is clean, structured, and optimized for model consumption.
- Serve as the go-to technical expert for all things data — pipelines, storage, security, and performance.
- Establish and enforce best practices for data engineering across the org.
Special Perks:
- Join a technology startup led by a 7x founder that has had successful exits.
- Be part of the foundational team and build products that solve massive problems and have huge upside.
- Have some skin in the game with a very aggressive equity component (along with strong base salary).
Must Have Skills:
What We’re Looking For:
- 6+ years of backend/data engineering experience, ideally in high-growth tech environments.
- Deep expertise in building and managing data pipelines (Airflow, Dagster, dbt, etc.).
- Strong Python skills (Pandas, FastAPI, PySpark).
- Experience with cloud data warehouses (Snowflake, BigQuery, Redshift) and storage systems (S3, GCS).
- Proficient with SQL and schema design.
- Hands-on experience with vector databases (e.g., Pinecone, Weaviate, FAISS).
- Familiarity with LLMs, embeddings, and building RAG pipelines.
- Strong understanding of data security and compliance best practices.
Nice to Have Skills:
Preferred Attributes:
- Experience in early-stage startups (0•1 product experience a big plus).
- Highly autonomous and comfortable in ambiguous environments.
- Excellent communication skills — able to work cross-functionally with product, design, and AI/ML.
- Proven ability to simplify complex data systems and make technical decisions with clarity.
- Experience building AWS data pipelines.