Summary
Overview
Work History
Education
Skills
Personal Information
Languages
Disclaimer
Certification
Timeline
Generic

NAGESH MOKRE

Pune

Summary

Overall 7.9 years of experience in IT and 7.9 years in Hadoop as a Hadoop Administrator Solid Knowledge of Hadoop framework and its Eco-System. Expertise in Hadoop architecture, design and development of Big Data platform including large clusters, Hadoop ecosystem projects Deployed Production grade as well as POC clusters on AWS. Designing and Development of High Availability Architecture to be hosted on AWS cloud Coordinated workflow on Virtual Private Cloud for selection of the IP address range, subnets, Route Tables and network gateways Configured and managing security groups and VPC Configuring Autoscaling, CloudWatch Alarms and SNS for Scaling and monitoring the Infrastructure. Expertise to install Hadoop and its related components in a multi-node cluster environment. Expertise in Implement new Hadoop hardware infrastructure Expertise in HDFS Architecture and Cluster concepts. Expertise in Hadoop Cluster capacity planning. Worked in 24x7 environment for production support in an on-call rotation. Expertise in Cluster Installation for Lab, Staging and Production environment Expertise in Hadoop job schedulers such as Fair scheduler and Capacity scheduler Knowledge of Cluster coordination services using Zoo Keeper. Expertise to install Hadoop and its related components in a multi-node cluster environment. Experience in setting up Open Source HortonWorks Data Platform through Amabri Experience in managing HDP upgrades through Ambari Experience in setting up security Should manage and clearly articulate in detail on managing HDP upgrades with rollback options Experience in configuring other Hadoop data distribution services like HBase, Hive, Impala etc. on a need basis Experience in configuring on demand Hadoop clusters Experience in AWS services like EC2 for installing and managing HDP clusters

Overview

8
8
years of professional experience
1
1
Certification

Work History

Hadoop Administrator

Wipro Technologies Ltd
04.2021 - Current
  • Handle 12 clusters and Monitoring all
  • Tuning by making changes to settings (e.g
  • HBase, Hive, NoSQL, etc.)
  • Responsible for commissioning and decommissioning of nodes from Clusters
  • Root Cause Analysis for all the failures (Systems, Ingestions, Jobs, Sync & Subscriptions) in production cluster
  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • Preparing and performing the installation of Hadoop software
  • Adding and configuring the nodes
  • Monitoring Cluster Health and Troubleshooting
  • Management of the meta data databases
  • Manage and review Backups
  • Moving data efficiently between clusters using Distributed Copy
  • Restore in case of physical data loss (e.g
  • Block corruptions)
  • Restore in case of project-specific requirements
  • Rebalancing of the HDFS
  • Regularly scheduling statistics runs and automating them
  • Quota administration, notification in case of space problems
  • Moving/shifting nodes/roles/services to other nodes
  • Performing cluster/node migrations (new hardware and/or new OS version and/or new Hadoop version) with OS or Hadoop tools or manually
  • Processing the events of Hadoop logs, taking measures, correcting errors, and involving the relevant teams if necessary
  • Tuning by making changes to settings (e.g
  • HBase, Hive, NoSQL, etc.).

Hadoop Administrator

Virtuous Soft Solutions Ltd Pune
07.2016 - 03.2021
  • Involved in cluster planning, capacity planning, and Hadoop phases
  • Deploying Cloudera Hadoop from scratch in AWS Virtual Private Cloud environment with EBS storage
  • Implementing security modules like enabling Kerberos integration with AD, RBAC using sentry with DB store/policy file, HDFS ACL, log/query Reduction
  • Responsible for CDH and Cloudera Manager upgrade, Commissioning and decommissioning nodes, enabling name node/resource manager HA
  • Manage services, roles, and rack awareness using Cloudera Manager
  • Troubleshooting application failures
  • Responsible for creating a Distcp workflow for securely copy data from S3 to HDFS & vice versa
  • YARN scheduler queue management using DRF, Log analysis, Troubleshooting yarn applications
  • Cluster User, quota, Trash, snapshot and snapshot policies management
  • Working on different services of AWS such as EC2, S3, EMR, VPC and Created Data pipelines from AWS S3 to HDFS and vice versa
  • Responsible for change management, incident management, report management
  • Responsible for attending the daily status call and change review meetings with the client
  • Develop and document best practices
  • Monitor Hadoop cluster connectivity and performance
  • Testing and bench-marking the new nodes
  • Managed CDH 5.13.3 cluster using Cloudera Manager and Linux RedHat 7.6
  • Configured various property files like core-site.xml, hdfs-site.xml, mapred- site.xml and hadoop-env.sh based upon the job requirement
  • Configuring YARN tuning
  • Successfully migrated Namenode from one node to another using cloudera Manager
  • Responsible for finding and troubleshooting Hadoop Errors
  • Working with data delivery teams to setup new Hadoop users
  • This job includes setting up Linux users, setting up Kerberos principals
  • Responsible for implementation and ongoing administration of Hadoop infrastructure
  • Preparing and performing the installation of Hadoop software
  • Adding and configuring the nodes
  • Monitoring Cluster Health and Troubleshooting
  • Management of the meta data databases
  • Manage and review Backups.

Education

Skills

  • Apache Hadoop
  • Hive
  • Sqoop
  • Pig
  • HBase
  • Flume
  • Cloudera
  • Windows
  • Ubuntu
  • Centos
  • MS-SQL
  • MySQL
  • Cloudera Manager
  • Nagios
  • Eclipse Memory Analyzer
  • AWS Certified Solutions Architect - Associate

Personal Information

  • Date of Birth: 06/14/92
  • Gender: Male
  • Nationality: Indian
  • Marital Status: Married

Languages

English, Hindi, Marathi

Disclaimer

I hereby declare that all the particulars given above are true and correct to the best of my knowledge.

Certification

  • AWS Certified Solutions Architect - Associate certificate (2020)

Timeline

Hadoop Administrator

Wipro Technologies Ltd
04.2021 - Current

Hadoop Administrator

Virtuous Soft Solutions Ltd Pune
07.2016 - 03.2021

NAGESH MOKRE