Must have good technical experience and should be able to understand and develop the modules and bring the task to closure on time.
At least 4+ years of development work experience in Hadoop programming (HDFS) using PySpark/Scala/Java with Hive-based Data warehouse projects, along with good Shell Scripting experience.
Able to understand the requirements and ingest the data into the Big Data platform using Spark with HIVE.
Should be able to write SQL scripts, including complex SQL scripts to process the data with SCD handling exposure.
Responsible for meeting quality and scope as per the project plans, must be able to meet stringent timelines.
Flexible to work as per regional timing.
Excellent communication and documentation skills.
Very good at team playing and flexible to work with the offshore team in different time zones based on project needs.
Willingness to obtain certification in Databricks Certified Developer: Apache Spark 3.X.
Nice to Have
Proactive, with good communication skills to articulate technical issues.
Exposure to Confluence/JIRA.
Ability to work independently; prior experience with databases like Oracle/SQL Server/ETL will be an added advantage.
At least 1+ years of Module Lead experience.
Hands-on experience in data analysis and debugging SQL issues.
Experience in performance tuning Hive QLs and cluster nodes, as well as functional testing, test planning, and execution.