Location: Singapore, Singapore
Thales is a global technology leader trusted by governments, institutions, and enterprises to tackle their most demanding challenges. From quantum applications and artificial intelligence to cybersecurity and 6G innovation, our solutions empower critical decisions rooted in human intelligence. Operating at the forefront of aerospace and space, cybersecurity and digital identity, we’re driven by a mission to build a future we can all trust.
In Singapore, Thales has been a trusted partner since 1973, originally focused on aerospace activities in the Asia‑Pacific region. With 2,000 employees across three local sites, we deliver cutting‑edge solutions across aerospace (including air traffic management), defence and security, and digital identity and cybersecurity sectors. Together, we’re shaping the future by enabling customers to make pivotal decisions that safeguard communities and power progress.
Whom We Are Looking For
We’re looking for a seasoned Data Architect to lead the design, adaptation and implementation of a Data Warehouse platform that serves as a central data repository for downstream processing to support the regulatory requirements, drive advanced analytics. In this role, you’ll architect scalable, unified data solutions that seamlessly combine the strengths of data lakes and warehouses—enabling advanced analytics, AI/ML applications, and secure, governed access to enterprise data. Join us to shape the foundation for next‑generation data‑driven innovation!
Responsibilities
- Owner of the Architecture of the Data Warehouse; ensures that systems architecture blueprints (e.g., architecture diagrams, tooling, technologies) remain current.
- Define data modeling standards for raw, curated and serving layers.
- Develop strategies for data ingestion, data storage, data cataloging, data governance, and secured data access (e.g., data‑at‑rest, data‑in‑transit).
- Develop strategies for data backups and data recovery based on SLA/RPO.
- Develop strategies for data consistency, data security (i.e., data‑at‑rest, data‑at‑transit) and data redundancy.
- Lead the design and implementation of ingestion pipelines, for structured and unstructured data, ensure that they meet the functional and non‑functional requirements.
- Lead the design of storage layers, metadata management, and data cataloging to ensure that they meet the functional and non‑functional requirements.
- Define ELT/ETL workflows and support automation with orchestration tools (e.g., Apache Spark, Apache Flink).
- Oversee deployment pipelines using CI/CD best practices for data‑oriented infrastructure.
- Ensure high availability, scalability and performance of the technical implementation (e.g., RMA analysis).
- Implement robust data security frameworks (i.e., encryption, data masking, and fine‑grained access controls).
- Establish IAM policies and secure data perimeters using cloud‑native and host‑based tools.
- Define and enforce data governance policies: data lineage, data quality and regulatory compliance.
- Collaborate with InfoSec and Compliance teams to run regular security audits, risk assessments and data assessments.
- Work in an agile, cross‑functional multinational team, actively engaging to support the success of the team.
Requirements
Education
- Bachelors in Computer Science or Information Technology
- Masters degree in Computer Science or Data Science, if applicable
Essential Skills / Experience
- Strong experience in designing, building highly available data platforms using MinIO as the data storage infrastructure, Data Replication / Partition strategies.
- Strong expertise in SQL, data modeling (SCD Type 1,2, 6), ETL/ELT design, and query performance tuning.
- Strong expertise in containerization and orchestration (i.e., Docker, Kaniko, Kubernetes)
- Strong expertise in design & development of ETL & ELT data pipelines (with structured or unstructured data), Change Data Capture.
- Strong expertise in distributed data processing technologies (Apache Spark 3.0, Apache Flink 2.0, Apache Iceberg, Trino, Apache Kafka)
- Hands‑on experience with scalable data storage platforms (e.g., Azure Data Lake Storage, MinIO S3).
- (Past / Present) Proficiency in implementing ETL/ELT that stores and retrieves data from object‑based data stores (e.g., MinIO) and relational data stores (e.g., PostgreSQL)
- Deep understanding of data security, encryption, IAM and compliance standards.
- Proficiency with integrating OTEL (OpenTelemetry) into Data Infrastructures
- Proficiency in programming languages in Java 8+, (e.g., Java 23.x), Kotlin 2.x
- Proficiency with Continuous Integration in using Git‑based protocols (e.g., Gitlab, Gitea).
- Proficiency with distributed source code management tools using Git‑based protocols (e.g., Gitlab, Gitea).
- Proficiency with using the Linux command line commands (e.g., Linux filesystem, Linux processes).
- Good communication skills in English
Desirable Skills / Experience
If you have the following desirable skills and relevant experiences, it would be an added advantage!
- Hands‑on experience with cloud‑native big data platforms (e.g., Azure Data Factory, Databricks).
- Working experiences with Python2/3, Scala2/3.
- Working experiences with working with Event‑Driven Architectures.
- Familiar with cloud‑native deployment strategies to cloud service providers (e.g., Azure Cloud, Amazon Cloud, Google Cloud).
- Familiar with the main cloud service models: Software as a Service, Platform as a Service and Infrastructure as a Service.
Essential / Desirable Traits
- Possess learning agility, flexibility and pro‑activity
- Comfortable with agile teamwork and user engagement
At Thales, we’re committed to fostering a workplace where respect, trust, collaboration, and passion drive everything we do. Here, you’ll feel empowered to bring your best self, thrive in a supportive culture, and love the work you do. Join us, and be part of a team reimagining technology to create solutions that truly make a difference – for a safer, greener, and more inclusive world.