Job Responsibilities
- Work across workstreams to support data requirements including reports and dashboards
- Analyze and perform data profiling to understand data patterns and discrepancies following Data Quality and Data Management processes
- Understand and follow best practices to design and develop the E2E Data Pipeline: data transformation, ingestion, processing, and surfacing of data for large-scale applications
- Develop data pipeline automation using Azure, AWS data platform and technologies stack, Databricks, Data Factory
- Understand business requirements to translate them into technical requirements that the system analysts and other technical team members can drive into the project design and delivery
- Analyze source data and perform data ingestion in both batch and real-time patterns via various methods; for example, file transfer, API, Data Streaming using Kafka and Spark Streaming
- Analyze and understand data processing and standardization requirements, develop ETL using Spark processing to transform data
- Understand data/reports and dashboards requirements, develop data export, data API, or data visualization using Power BI, Tableau, or other visualization tools
Qualifications
We are looking for experience and qualifications in the following:
- Bachelor’s degree in Computer Science, Computer Engineer, IT, or related fields
- Minimum of 4 years’ experience in Data Engineering fields
- Data Engineering skills: Python, SQL, Spark, Cloud Architect, Data & Solution Architect, API, Databricks, Azure, AWS
- Data Visualization skills: Power BI (or other visualization tools), DAX programming, API, Data Model, SQL, Story Telling and wireframe design
- Business Analyst skills: business knowledge, data profiling, basic data model design, data analysis, requirement analysis, SQL programing
- Basic knowledge in Data Lake/Data Warehousing/ Big data tools, Apache Spark, RDBMS and NoSQL, Knowledge Graph
- Experience working in a Singapore public sector, client-facing/consulting environment is a plus
- Team player, analytical and problem-solving skills
Mandatory Skills: PySpark, Databricks