• Hadoop Developer/Admin

Industry IT
Location Maharashtra Pune
Experience Range 3 - 5 Years
Qualification BE
Not active

Functional IT Software - Application Programming / Maintenance
Job Description
About Us
“Quess IT Staffing is India’s largest IT staffing company with over 20 years of experience in staffing IT professionals in 300+ companies across levels and skillsets. Our 10,000+ associates deployed in 80+ cities and towns are proficient in over 500 technological skills. Our associates help enable cutting edge solutions some of the biggest names across industried. Quess IT Staffing is a division of Quess Corp Limited, India’s leading business services provider and largest domestic private sector employer. Quess Corp Limited is - ‘A Great Place to Work’ certified – a testament to our excellent culture, people, and processes.”
About Company
https://itstaffing.quesscorp.com/
Roles and Responsibility

·         Providing Administrative Support for Customers on Hadoop platforms. Typically, these customers may have 24/7 contracts, and the successful applicant must be prepared to work in shifts and also be on-call to support customer site/s per contractual obligations.

·        The Hadoop Administrator manages and controls the Hadoop System environment for Teradata customers.

·        The Hadoop Administrator requires specific technical knowledge about the administration and control of the Hadoop System, including the associated operating system, related tools, network, and hardware.

 

Desired Competencies/Experience:

·         Mandatory Skills :

·         Minimum experience of 3-6 years in Managing and Supporting large scale Production Hadoop environments (configuration management, monitoring, and performance tuning) in any of   the Hadoop distributions (Apache, Hortonworks, Cloudera, MapR, IBM BigInsights, Pivotal HD).

·         3-6 years of experience in Scripting Language (Linux, SQL, Python). Should be proficient in shell scripting.

·         3-6 years of experience on Administrative activities likes- Administration, maintenance, control, and optimization of Hadoop capacity, security, configuration, process scheduling, and errors.

·         Management of data, users, and job execution on the Hadoop System.

·         Experience in Backup, Archival and Recovery (BAR) and High availability (HA).

·         Plan for and support hardware and software installation and upgrades.

·         3-6 years of Experience in Hadoop Monitoring tools (Cloudera Manager, and Ambari, Nagios, Ganglia etc).

·         Experience may include (but is not limited to) build and support including design, configuration, installation (upgrade), monitoring and performance tuning of any of the Hadoop   distributions.

·         Hadoop software installation and upgrades.

·         Experience of workload / performance management.

·         Automation - experience in CI / CD (Continuous Integration / Deployment) Jenkins, Ansible, Terraform, Puppet, Chef.

·         Implementing standards and best practices to manage and support data platforms as per distribution.

·         Proficiency in Hive internals (including HCatalog), SQOOP, Pig, Oozie and Flume/Kafka.

·         Experience in MySQL & PostgreSQL databases.

·         ITIL Knowledge.

 

·         Preferred Skills:

·         Experience with DR (Disaster Recovery) strategies and principles.

·         Development or administration on NoSQL technologies like Hbase, MongoDB, Cassandra, Accumulo, etc.

·         Development or administration on Web or cloud platforms like Amazon S3, EC2, Redshift, Rackspace, OpenShift etc.

·         Development/scripting experience on Configuration management and provisioning tools e.g. Puppet, Chef.

·         Web/Application Server & SOA administration (Tomcat, JBoss, etc.)

·         Development, Implementation or deployment experience on the Hadoop ecosystem (HDFS, MapReduce, Hive, Hbase).

·         Experience on any one of the following will be an added advantage:

·         Hadoop integration with large scale distributed data platforms like Teradata, Teradata Aster, Vertica, Greenplum, Netezza, DB2, Oracle, etc.

·         Proficiency with at least one of the following: Java, Python, Perl, Ruby, C or Web-related development

·         Knowledge of Business Intelligence and/or Data Integration (ETL) operations delivery techniques, processes, methodologies

·         Exposure to tools data acquisition, transformation & integration tools like Talend, Informatica, etc. & BI tools like Tableau, Pentaho, etc.

·         Linux Administrator certified.

 

 

 

A+| A| A-