Location: Farmington Hills, MI
Duration: 6+ Months
Minimum Skills required:
- 4+ years of development experience with Hadoop ecosystem (Spark, Scala, Oozie, Pig, Hive, HDFS, MapReduce) and/or NoSQL technologies such as Cassandra, MongoDB with experience on real-time & stream processing systems. POC experience or Training won't be considered
- Excellent knowledge of Core Java or UNIX shell script or PL/SQL stored procedures is required
- Should have knowledge in different Hadoop Distributions like CDH 4/5, Hortonworks, MapR, IBM Big Insights.
- Strong foundational knowledge and experience with a range of Big data components such as Hadoop/Yarn, HDFS, MapReduce, Oozie, Falcon, Pig, Hive, Zookeeper, Sqoop and Flume
- Develop MapReduce programs or Hadoop streaming.
- Develop Pig scripts/Hive QL for analyzing all semi-structured/unstructured/structured data flows.
- Should have knowledge of table definitions, file formats, UDF, Data Layout ( Partitions & Buckets), Debugging & performance optimizations.
- Excellent oral and written communication skills.