Location: Farmington Hills, MI
Duration: 6+ Months
Minimum Skills required:
- – 4+ years of development experience with Hadoop eco-system ( Spark, Scala, Oozie, , Pig, Hive ,HDFS, MapReduce) and/or NoSQL technologies such as Cassandra, MongoDB with experience on real-time & stream processing systems. POC experience or Training won’t be considered
- – Excellent knowledge of Core Java or UNIX shell script or PL/SQL stored procedures is required
- – Should have knowledge in different Hadoop Distributions like CDH 4 / 5, Hortonworks, MapR, IBM Big Insights.
- – Strong foundational knowledge and experience with a range of Big data components such as Hadoop/Yarn, HDFS,MapReduce,Oozie, Falcon, Pig, Hive, Zookeeper, Sqoop and Flume
- – Develop MapReduce programs or Hadoop streaming.
- – Develop Pig scripts/Hive QL for analyzing all semi-structured/unstructured/structured data flows.
- – Should have knowledge of Table definitions, file formats, UDF , Data Layout ( Partitions & Buckets), Debugging & performance optimizations.
- – Excellent oral and written communication skills.