Hello,

Here is new opening please review and send your resume as soon as possible.

*Title: Hadoop admin*
*Location: Bellevue, WA*
*Duration:6+ Months*
*Experience:7+ Years*
*Visa Copy Required*

*Hortonworks  is mandatory for this position*


*Job Description:-*


   -  Responsible for implementation and ongoing administration of Hadoop
    infrastructure.
   -  Aligning with the systems engineering team to propose and deploy new
   hardware and software environments required for Hadoop and to expand
   existing environments.
   -  Working with data delivery teams to setup new Hadoop users. This job
   includes setting up Linux users, setting up Kerberos principals and
   testing HDFS, Hive, Pig and MapReduce access for the new users.
   -  Cluster maintenance as well as creation and removal of nodes using
   tools like Ganglia, Nagios etc.
   - Performance tuning of Hadoop clusters and Hadoop MapReduce routines.
   -  Screen Hadoop cluster job performances and capacity planning
   -  Monitor Hadoop cluster connectivity and security
   -  Manage and review Hadoop log files.
   -  File system management and monitoring.
   -  HDFS support and maintenance.
   -  Diligently teaming with the infrastructure, network, database,
   application and development  teams to guarantee high data quality and
   availability.
   -  Collaborating with application teams to install operating system and
   Hadoop updates, patches, version upgrades when required.
   -  Upgrade Hadoop Hortonworks versions from 2.1x to higher
   -  Backup and disaster recovery
   -  Point of Contact for Vendor escalation
   - Skills Required:
   - General operational expertise such as good troubleshooting skills,
   understanding of system’s capacity, bottlenecks, basics of memory, CPU, OS,
   storage, and networks.
   -  Hadoop skills like HBase, Hive, Pig, Mahout, etc.
   -  Hortonworks upgrade experience from 2.1x to higher version
   - The most essential requirements are: They should be able to deploy
   Hadoop cluster, add and remove nodes, keep track of jobs, monitor
   critical parts of the cluster, configure name-node high availability,
   schedule and configure it and take backups.
   -  Good knowledge of Linux as Hadoop runs on Linux.
   - Familiarity with open source configuration management and deployment
   tools such as Puppet or Chef and Linux scripting.
   -  Knowledge of Troubleshooting Core Java Applications is a plus.

-- 
You received this message because you are subscribed to the Google Groups 
"International SAP Projects" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to international-sap-projects+unsubscr...@googlegroups.com.
To post to this group, send email to 
international-sap-projects@googlegroups.com.
Visit this group at https://groups.google.com/group/international-sap-projects.
For more options, visit https://groups.google.com/d/optout.

Reply via email to