About the Role
We're looking for a hands-on Principal Data Engineer to lead the design and build of our AI/ML data infrastructure. You'll architect scalable pipelines and data systems that support everything from model training to real-time inference. This is a key technical role that connects the dots between raw data, product development, and AI innovation across the company.
Key Responsibilities
- AI Data Infrastructure: Design and own our data architecture for AI and ML, from data collection to feature engineering and model deployment.
- Feature Store & Pipelines: Build a centralized feature store and develop high-throughput pipelines for training, batch, and real-time inference.
- Scalable Systems: Create streaming and batch data systems using tools like Kafka, Kinesis, or Pulsar.
- Data Modelling: Design and evolve our Data Warehouse for both analytics and ML workflows.
- Cloud & Serverless: Drive a cloud-native, serverless-first architecture using AWS, GCP, or Azure services.
- Data Quality & Governance: Implement data lineage, monitoring, security, and privacy protocols that support responsible AI.
- Technical Leadership: Mentor engineers and data scientists, champion best practices, and drive a data-driven culture.
- Cross-Team Collaboration: Work closely with product, engineering, and business teams to turn business needs into scalable data solutions.
Who You Are
- 10+ years of experience in data engineering or infrastructure, with deep hands-on experience in high-scale production systems.
- Strong knowledge of cloud platforms and services (AWS, GCP, or Azure) including data tools like S3, Redshift, BigQuery, Glue, or Dataflow.
- Proficient in Python (or similar), SQL, and NoSQL databases.
- Solid experience with streaming systems (Kafka, Pulsar, etc.).
- Experience building AI/ML data workflows and deploying pipelines for training/inference cycles.
- Strong system design and data modeling skills for both operational and analytical use cases.
- Proven ability to define and uphold data quality, security, and governance.
- Comfortable leading projects and mentoring teams.
Nice to Have
- Experience with MLOps tools and CI/CD (e.g., Terraform, CloudFormation, Docker, Kubernetes).
- Familiarity with NLP or Computer Vision data pipelines.
- Hands-on experience with feature store frameworks (Feast, Tecton, or custom).
- Comfortable communicating technical topics to non-technical stakeholders.
Why Join Us
- Help build and scale the next generation of data infrastructure powering AI in retail and trade.
- Work in a fast-moving, multicultural team backed by Temasek.
- Competitive salary and performance-based compensation.
- Flexible remote/hybrid work options.