Hadoop Administrator Charlotte, NC / New Jersey. Contract. Tasks and Major Responsibilities include: Installation, configuration and Upgrading Cloudera distribution of Hadoop Responsible for implementation and on-going administration of Hadoop infrastructure Create and document best practices for Hadoop and Big data environment Experience working on hadoop ecosystem components like hdfs, hive, map-reduce, yarn, impala, spark, Sqoop, HBase, Sentry, Hue and oozie Experience working on Hadoop security aspects including Kerberos setup, RBAC authorization using Apache Sentry Working with data delivery teams to setup new Hadoop users. (job includes setting up Linux users, setting up Kerberos principals and testing HDFS, Hive, Pig and Map Reduce access for the new users File system management and cluster monitoring using Cloudera Manager Performance tuning of Hadoop clusters and Hadoop MapReduce routines Strong troubleshooting skills involving map reduce, yarn, sqoop job failure and its resolution Backup and disaster recovery solution for Hadoop cluster Willing to participate and learn new third party technologies to make Hadoop environment efficient and secure Experience working on Unix operating system who can efficiently handle system administration tasks related to Hadoop cluster Knowledge or experience working on no-sql databases like Hbase, Cassandra, Mongodb Troubleshooting connectivity issues between BI tools like Datameer, Sas and Tableau and Hadoop cluster Point of contact for vendor escalation Should be available for 24*7 Hadoop support issues Participate in new data product or new technology evaluation; manage certification process. Evaluate and implement new initiatives on technology and process improvements. Interact with Security Engineering to design solutions, tools, testing and validation for controls
Technical Skills: Candidate will have 2 to 3 year experience with Hadoop data stores/cluster administration and 5 to 8 years relational database experience. Strong Hadoop cluster administration expertise; Understanding of internals Excellent performance and tuning skills of large workload inside Hadoop cluster Strong Partitioning knowledge ( data sharding concepts) Scripting Skills – Shell and Python Experience in upgrading Cloudera Hadoop distributions Experience in performance tuning and troubleshooting - drill down approach with O/S, database and application - End to End Application conn Familiarity with NoSQL data stores (MongoDB / Cassandra/HBase) Cloud Architecture (Public and Private clouds) – AWS , AZURE familiarity Prior experience of administration of Oracle or any other relational database Desired Skills (Preferred not required): Scripting with Pig Proficient in using Microsoft Office 2010 (Word, Excel, PowerPoint) Automation Tools – Puppet , CFEngine Familiarity with Solr, Spark is preferred Data Movement Tools such as Data Stage, Informatica, Sqoop etc Thanks & Regards, Syed Kumail. Email : kumai...@scadea.net 908-636-2727 Extn 2240 -- You received this message because you are subscribed to the Google Groups "OracleD2K" group. To unsubscribe from this group and stop receiving emails from it, send an email to oracled2k+unsubscr...@googlegroups.com. To post to this group, send email to oracled2k@googlegroups.com. Visit this group at https://groups.google.com/group/oracled2k. For more options, visit https://groups.google.com/d/optout.