RAKESH

Lead Big Data Administrator
 Location  Gurgaon
 Total Views  89
 Shortlist 0
 Member since  30+ Days ago
 
Candidate Information
  • Experience  10 Year
  • Hourly Rate $20
  • Availability Immediate
  • Work From Any
  • Category Information Technology & Services
  • Last Active On October 23, 2021
Key Skills
Microsoft Azure Admin shell Scripting Jenkins Hadoop TFS Eclipse
Summary

Hi,


I feel that my skills and experience are a great fit for this position. 

Please feel free to contact me to arrange an interview. I look forward to learning more about this opportunity.

Please find the attached resume.


Experience Summary: - I have 10 years of working experience in mix of profile like Hadoop Administrator and Software Testing.

I have 4 years of work experience as a Hadoop Administrator.


Big Data Exposure:-

*Installation and Verify in Hadoop logic installation, configuration, supporting and managing Hadoop Clusters using Apache Hadoop.

*Verify backup configuration and recovery from a Namenode failure.

*Verify Installation of various Hadoop Ecosystems and Hadoop Daemons.

*Checking system health by Heartbeat mechanism. 

*Good experience on Design, configure and manage the backup and disaster recovery for Hadoop data.

*Hands on experience in analyzing log files for Hadoop and ecosystem services and finding  root cause. 

*As an administrator verify, cluster maintenance, troubleshooting, monitoring and followed proper backup & Recovery strategies.

*Experience in HDFS data storage and support for running map-reduce jobs. 

*Verify installing and configuring Hadoop eco- system like Sqoop, Pig, Hive, Hbase, Flume, Oozie, and Kafka.

*Configured various property files like core-site.xml, hdfs-site.xml, mapred-site.xml, yarn-site.xml based upon the job requirement.

*Importing and exporting data into HDFS using Sqoop.

*Importing and exporting data into hdfs to local, local to local.

*Good working Knowledge in Hadoop security like Kerberos and sentry.

*Experienced in Cloudera installation, configuration and deployment on Linux distribution.

*Commissioning and decommissioning of nodes as require.

*Managing and monitoring Hadoop services like Namenode, Datanode & Yarn

*Performance tuning, and solving Hadoop issues using CLI or by WebUI

*Troubleshooting Hadoop cluster runtime errors and ensuring that they do not occur again.

*Accountable for storage and volume management of Hadoop clusters.

*Ensuring that the Hadoop cluster is up and running all the time (High availability, big data cluster etc.)

*Evaluation of Hadoop infrastructure requirements and design/deploy solutions.

*Backup and recovery task by creating snapshots policies, backup schedules and recovery from node failure.

*Responsible for Configuring Alerts for different types of services which is running in  Hadoop Ecosystem.

*Moving data from one cluster to another.

 

Warms Regards

Rakesh Dubey

8744855286