Data Engineer (1-year renewable contract)
EVOLUTION RECRUITMENT SOLUTIONS PTE. LTD.
Singapore
On-site
SGD 70,000 - 100,000
Full time
Job summary
A leading recruitment solutions company in Singapore seeks a skilled data professional to oversee data governance and manage high-performance data processing pipelines. The ideal candidate should have extensive experience with big data tools, Master Data Management platforms, and strong teamwork skills. This role offers an opportunity to collaborate with cross-functional teams and drive data-insight initiatives.
Qualifications
- Proven experience building and operating large-scale data lakes and data warehouses.
- Strong knowledge of Hadoop ecosystem and big data tools.
- Hands-on experience with Master Data Management (MDM) platforms.
Responsibilities
- Create and manage a single master record for each business entity.
- Implement data governance processes.
- Build and maintain robust data processing pipelines.
Skills
Hadoop ecosystem
Spark
Kafka
MDM platforms
SQL optimization
Python
Java
Scala
Tools
Key Responsibilities
- Create and manage a single master record for each business entity to ensure data consistency, accuracy, and reliability.
- Implement data governance processes, including quality management, profiling, remediation, and automated data lineage.
- Build and maintain robust, high-performance data processing pipelines across cloud, private data centers, and hybrid ecosystems.
- Assemble and process large, complex datasets from diverse data sources.
- Collaborate with Data Scientists, ML Engineers, Business Analysts, and other stakeholders to deliver actionable insights and improve business performance.
- Develop, deploy, and maintain microservices, REST APIs, and reporting services.
- Design and automate internal processes to streamline workflows, optimize data delivery, and scale infrastructure.
- Troubleshoot and analyze large-scale distributed systems to ensure reliability and performance.
- Work closely with cross-functional teams in a dynamic and fast-paced environment.
Key Requirements
- Proven experience building and operating large-scale data lakes and data warehouses.
- Strong knowledge of Hadoop ecosystem and big data tools, including Spark and Kafka.
- Hands-on experience with Master Data Management (MDM) platforms such as Informatica MDM, Talend Data Catalog, Semarchy xDM, IBM PIM & IKC, or Profisee.
- Familiarity with MDM processes (golden record creation, survivorship, reconciliation, enrichment, quality).
- Experience in data governance, including data profiling, remediation, and automated lineage.
- Knowledge of stream-processing systems (e.g., Spark Streaming).
- Proficiency with cloud services (Azure, GCP, AWS) and platforms like Delta Lake, Databricks.
- Advanced experience with relational and NoSQL databases (Hive, HBase, Postgres).
- Strong SQL optimization skills.
- Proficiency in programming/scripting languages such as Python, Java, Scala.
- Proven ability to manipulate, process, and extract value from large and disconnected datasets.
- Familiarity with modern development practices (Scrum, TDD, CI/CD, code reviews).
- Strong teamwork skills, with proven success in cross-functional collaboration.
Preferred Skills
- Experience with ETL tools (Talend Big Data, Azure Data Factory, etc.).
- Exposure to Cloudera Data Platform.
- Experience with metadata management and data governance tools.
- Previous involvement in Master Data Management projects.