Summary
Overview
Work History
Education
Skills
Certification
Projectshandled
Personaldossier
Careerpath
Timeline
Intern
Venkatesh M R

Venkatesh M R

Bangalore

Summary

Driven Technical Architect with 12+ years of success in conceptualizing technical solutions and system development environments. Focused on working closely with clients to determine technical project objectives, specifications and solutions. Detail-oriented in coding and testing the functionality of deliverables.

Overview

2025
2025
years of professional experience
2008
2008
years of post-secondary education
1
1
Certification

Work History

Technical Architect

NucleusTeq
5 2024 - Current
  • Participated in customer presentations as a Technical Architect to demonstrate capabilities in Big Data and AWS EMR Cloud technologies
  • Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements
  • Participated in customer PoC deliveries, showcasing Big Data solutions on on-prem, AWS EMR and demonstrating business value
  • Resolved an age old S3 storage issue for Spark job intermediate failures and created a POC to withhold Shuffle data while AWS Spot instances decommissioning, which saved huge cost
  • Responsible for designing and developing Airflow jobs to process CRM batch data
  • Performance tuning and optimization to ensure efficient resource utilization
  • Troubleshooting and resolution of performance-related issues
  • Automated intermediate data processing from S3 using AWS Lambda
  • Worked closely with stakeholders to develop standard operating procedures for effective platform management, and participated in customer-facing presentations as a Technical Architect
  • Developed reusable data pipeline templates and accelerator components to enhance efficiency in implementing Big Data solutions, reducing deployment time for client projects

Staff Tech Engineer

Alation
05.2023 - 02.2024
  • Expertise in designing, implementing, and managing Big Data solutions using technologies such as Hadoop, Spark, Kafka, Flink, etc
  • Hands-on experience with major Hadoop distributions (Cloudera, Hortonworks) and related ecosystem components
  • Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements
  • Participated in customer PoC deliveries, showcasing Big Data solutions on on-prem, AWS EMR and demonstrating business value
  • Installation, configuration, and management of Hadoop clusters, including CDP, HDP, CDH or other distributions
  • Expertise in Hadoop ecosystem components such as HDFS, MapReduce, YARN, Zookeeper, Hive, Spark, Kerberos, Ranger., etc
  • Performance tuning and optimization to ensure efficient resource utilization
  • Troubleshooting and resolution of performance-related issues
  • Integration of various data sources into the Hadoop ecosystem
  • Configuration and management of security protocols, including SSL/TLS and data encryption
  • Experience with deploying and managing Hadoop clusters on cloud platforms like AWS, Azure
  • Extensive expertise implementing Alation Data Governance software from V2021.x to V2023.x
  • Working knowledge of essential Alation components such as Data Dictionary, Reference Data, Business Glossary, Articles, and Alation Docker service Agents for various data sources
  • Working with new customers after sales to formalize the Alation product process and integrate data sources into the Alation platform
  • Identifying product defects and reporting them to developers
  • In charge of testing new component enhancement features
  • Worked with the engineering teams to get the Hadoop's HDFS data catalog into the Alation
  • Collaborated with Solution Engineers on the POC calls for data source component specific to bridging the gap of Alation integration
  • Collaborating with engineering teams to improve Alation connections for Oracle single instance and RAC instances databases
  • Collaborating with engineering teams on the Hive vs
  • Impala OCF (Open Connector Framework) project

Solutions Engineer

Acceldata Technology PVT LTD
04.2021 - 05.2023
  • Contributed to a new Hadoop build flavor known as ODP, here involved in building and testing Hadoop services (Open Data Platform)
  • Apache Hue and Apache Impala were built and integrated with ODP using mpack
  • Performed in-place data migration from CDH/HDP to ODP platform without any data loss
  • Worked closely with infrastructure, network, database and application teams at the client side to make team understand the ODP migration, get their support when required and continued further support post migration
  • Provided Vendor based support to the multiple clients
  • Architected Big Data platforms on on-prem for a banking customer
  • Implementation and support of the Enterprise Hadoop Environment
  • Capacity planning and performance tuning
  • Actively contributed to the enhancement of CDH/CDP/HDP/ODP Hadoop monitoring capabilities on a tool called Pulse
  • Work closely with infrastructure, network, database and application teams at the client side
  • Monitoring and managing storage that is used by Hadoop system
  • Configuring security on Hadoop cluster
  • Worked with the multiple clients to install the Hadoop monitoring tool called Pulse, demonstrated the benefits of the Pulse and its daily usage

Technology Services Specialist

IBM
04.2017 - 02.2021
  • Managed and built several Hadoop clusters in production, development, Disaster Recovery environments
  • Involved in Hadoop CDH patching and CDH upgrade activities
  • Managed and built several Hadoop clusters in production, development, Disaster Recovery environments
  • Work with engineering software developers to investigate problems and make changes to the Hadoop environment and associated applications
  • Expertise in recommending hardware configuration for Hadoop cluster
  • Installing, Upgrading and Managing Hadoop Cluster on Cloudera distribution
  • Trouble shooting many cloud related issues such as Data Node down, Network failure and any OS related memory issues
  • Built and configured log data loading into HDFS using Flume
  • Helped developers in configuring Sqoop Importing and exporting data into HDFS and Hive tables
  • Provisioning, installing, configuring, monitoring, and maintaining HDFS, Yarn, Sqoop, Hive, Impala, HBase, Hue, Oozie, Spark, Sentry, etc
  • Recovering from node failures and troubleshooting common Hadoop cluster issues
  • Secured the cluster by enabling the Kerberos AD integration with cloudera manager and automated the TGT renewal and ticket generation for all the service accounts
  • Designed the users and groups creation based on the data loads and data tracks
  • Enabled the granular data access for those users by integrating sentry with the cloudera
  • Experience on decommissioning failed nodes and commissioning new nodes as the cluster grows and to accommodate more data on HDFS
  • Designed end to end backup disaster recovery for production cluster and make sure the replication should happened on time with the minimal data loss
  • Experience in enabling High Availability to avoid any data loss or cluster down time
  • Hands on experience in cluster upgrade and patching without any data loss and with proper backup plans
  • Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes
  • Supporting Hadoop developers and assisting in optimization of mapreduce jobs, Sqoop-Hive Scripts, and HBase ingest, Kudu and Spark job failures if required
  • System/cluster configuration and health check-up
  • Continuous monitoring and managing the Hadoop cluster through Cloudera UI
  • Resolving tickets submitted by users, troubleshoot the error documenting, resolving the errors
  • Performed HDFS cluster support and maintenance tasks like Adding and Removing Nodes without any effect to running jobs and data
  • Experience in setting up high available to Hive Metastore, Hue
  • Environment: Hadoop HDFS, Zookeeper, Map Reduce, Sqoop, Hive, Impala, HBase, Kudu, Flume, Nagios, Sqoop, Linux, Cloudera, Mysql DB, Postgresql DB, Oracle DB
  • Proven results-oriented person with a focus on delivery
  • Conducting training sessions to new joiners and also on any new tool release

IT Analyst

TCS
02.2015 - 02.2017
  • Installed and configured Hadoop, YARN, MapReduce, Flume, HDFS (Hadoop Distributed File System)
  • Experience in installing Cloudera CDH stack of Hadoop and other ecosystem components
  • Managed Hadoop clusters: setup, install, monitor, maintain
  • Hands on experience in managing and monitoring the cluster using Cloudera UI dashboard
  • Conducting root cause analysis and resolve production problems and data issues
  • Proactively involved in ongoing maintenance, support and improvements in Hadoop cluster
  • Document and manage failure/recovery (loss of name node, loss of data node, replacement of HW or node)
  • Involved in Minor and Major Release work activities
  • Executed tasks for upgrading cluster on the staging platform before doing it on production cluster
  • Installation and Administration of a Hadoop cluster
  • Experience in installing and configuring of HBase
  • Monitor Hadoop Name node Health status, number of Task trackers running, number of Data nodes running
  • Adding nodes into the clusters & removing nodes from the cluster for maintenance
  • Experienced in defining job flows
  • Experienced in managing and reviewing Hadoop log files
  • Environment: Hadoop, YARN, HDFS, HBASE, Flume, MapReduce, Sqoop, Zoo Keeper, Oozie, NoSQL, Oracle and Linux

DBA

Genpact India
06.2011 - 02.2015
  • Supported and assisted a team of DBA's in migration of databases from Oracle 9i to 10g to 11g
  • The role involved in installation, configuration & up-gradation of oracle databases using exp/imp utilities
  • Setup/configured oracle RMAN for database backup and recovery
  • Applied different recovery procedures for complete and incomplete recoveries
  • Cloned the instances from production to Test, Development instance to test & simulate few technical challenges, coordinate with testing and development teams
  • Performed Database Cloning and refresh on test and develop environment

Education

BSc -

Sri Venkateswara University

12th - Board of Intermediate Education

Narayana Junior College

10th - Board of Secondary Education

Margadarsi E M H School

Skills

Hadoop Architect

Certification

Certified AWS Solutions Architect Associate

Projectshandled

NucleusTeq, 05/2024 - Present, Participated in customer presentations as a Technical Architect to demonstrate capabilities in Big Data and AWS EMR Cloud technologies., Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements., Alation, 05/2023 - 02/2024, Expertise in designing, implementing, and managing Big Data solutions using technologies such as Hadoop, Spark, Kafka, Flink, etc., Hands-on experience with major Hadoop distributions (Cloudera, Hortonworks) and related ecosystem components.

Personaldossier

Currently seeking a challenging position where I can continue to uphold high performance standards.

Careerpath

  • NucleusTeq, 05/2024 - Present
  • Alation, 05/2023 - 02/2024
  • Acceldata Technology PVT LTD, 04/2021 - 05/2023
  • IBM, 04/2017 - 02/2021
  • TCS, 09/2015 - 02/2017
  • Future Focus PVT LTD. Deployed to client place TCS, 02/2015 - 09/2015
  • Genpact, Bangalore, 06/2011 - 02/2015

Timeline

Staff Tech Engineer

Alation
05.2023 - 02.2024

Solutions Engineer

Acceldata Technology PVT LTD
04.2021 - 05.2023

Technology Services Specialist

IBM
04.2017 - 02.2021

IT Analyst

TCS
02.2015 - 02.2017

DBA

Genpact India
06.2011 - 02.2015

12th - Board of Intermediate Education

Narayana Junior College

10th - Board of Secondary Education

Margadarsi E M H School

Technical Architect

NucleusTeq
5 2024 - Current

BSc -

Sri Venkateswara University
Venkatesh M R