Job Type : W2

Experience : 1-2 yrs

Location : 10-Oct-2017

Posted Date :

Description :

Description :

Hadoop development engineer with good Knowledge on Hadoop echo systems HDFS (Cloudera preferred), Map Reduce, Pig, Sqoop and Hive for scalability, distributed computing and performance computing.

 

Responsibilities:

  • Analyze Hadoop cluster using different bigdata analytic tools including MapReduce, Hive, Sqoop, Zookeeper, Spark, Spark-streaming, Kafka, Phoenix-HBase.
  • Translate functional and technical requirements into a details programs unning on Hadoop and Spark.
  • Develop Logic in Spark and Spark Streaming using Scala and HQL for fast processing of the data according to the business requirements.
  • Design Multiple POC’s using Spark and Scala and deploy on cluster, compare the performance of Sparksql, with hive.
  • Develop and create Hive tables to store high volumes of data receiving from different data sources into a single framework and, work on performance tuning.
  • Implementation of spark jobs with parametrization and incremental data storage.
  • Write Shell Scripts for Automation of job to run in all environments for various checks to monitor the cluster health.
  • Work with team in analyzing large data sets to provide strategic direction on implementation.
  • Maintain documentation of the code development and deliver detailed documentation for design and production support.
  • Monitor Production Jobs, provide on call support for the project and resolve issues, tickets raised with in the business.

Requirements:

  • Minimum of Bachelor’s degree in Computer Science, Information Systems, or related field or its Equivalent.
  • Good Knowledge in HDFS (Cloudera preferred), Map Reduce, Pig, Sqoop, Hbase and Hive
  • Candidate should have a good overall understanding of Data and Analytics.
  • Should have scripting skills in Linux environment.
  • Good communication skills.