Driven Technical Architect with 12+ years of success in conceptualizing technical solutions and system development environments. Focused on working closely with clients to determine technical project objectives, specifications and solutions. Detail-oriented in coding and testing the functionality of deliverables.
Overview
2025
2025
years of professional experience
2008
2008
years of post-secondary education
1
1
Certification
Work History
Technical Architect
NucleusTeq
5 2024 - Current
Participated in customer presentations as a Technical Architect to demonstrate capabilities in Big Data and AWS EMR Cloud technologies
Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements
Participated in customer PoC deliveries, showcasing Big Data solutions on on-prem, AWS EMR and demonstrating business value
Resolved an age old S3 storage issue for Spark job intermediate failures and created a POC to withhold Shuffle data while AWS Spot instances decommissioning, which saved huge cost
Responsible for designing and developing Airflow jobs to process CRM batch data
Performance tuning and optimization to ensure efficient resource utilization
Troubleshooting and resolution of performance-related issues
Automated intermediate data processing from S3 using AWS Lambda
Worked closely with stakeholders to develop standard operating procedures for effective platform management, and participated in customer-facing presentations as a Technical Architect
Developed reusable data pipeline templates and accelerator components to enhance efficiency in implementing Big Data solutions, reducing deployment time for client projects
Staff Tech Engineer
Alation
05.2023 - 02.2024
Expertise in designing, implementing, and managing Big Data solutions using technologies such as Hadoop, Spark, Kafka, Flink, etc
Hands-on experience with major Hadoop distributions (Cloudera, Hortonworks) and related ecosystem components
Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements
Participated in customer PoC deliveries, showcasing Big Data solutions on on-prem, AWS EMR and demonstrating business value
Installation, configuration, and management of Hadoop clusters, including CDP, HDP, CDH or other distributions
Expertise in Hadoop ecosystem components such as HDFS, MapReduce, YARN, Zookeeper, Hive, Spark, Kerberos, Ranger., etc
Performance tuning and optimization to ensure efficient resource utilization
Troubleshooting and resolution of performance-related issues
Integration of various data sources into the Hadoop ecosystem
Configuration and management of security protocols, including SSL/TLS and data encryption
Experience with deploying and managing Hadoop clusters on cloud platforms like AWS, Azure
Extensive expertise implementing Alation Data Governance software from V2021.x to V2023.x
Working knowledge of essential Alation components such as Data Dictionary, Reference Data, Business Glossary, Articles, and Alation Docker service Agents for various data sources
Working with new customers after sales to formalize the Alation product process and integrate data sources into the Alation platform
Identifying product defects and reporting them to developers
In charge of testing new component enhancement features
Worked with the engineering teams to get the Hadoop's HDFS data catalog into the Alation
Collaborated with Solution Engineers on the POC calls for data source component specific to bridging the gap of Alation integration
Collaborating with engineering teams to improve Alation connections for Oracle single instance and RAC instances databases
Collaborating with engineering teams on the Hive vs
Impala OCF (Open Connector Framework) project
Solutions Engineer
Acceldata Technology PVT LTD
04.2021 - 05.2023
Contributed to a new Hadoop build flavor known as ODP, here involved in building and testing Hadoop services (Open Data Platform)
Apache Hue and Apache Impala were built and integrated with ODP using mpack
Performed in-place data migration from CDH/HDP to ODP platform without any data loss
Worked closely with infrastructure, network, database and application teams at the client side to make team understand the ODP migration, get their support when required and continued further support post migration
Provided Vendor based support to the multiple clients
Architected Big Data platforms on on-prem for a banking customer
Implementation and support of the Enterprise Hadoop Environment
Capacity planning and performance tuning
Actively contributed to the enhancement of CDH/CDP/HDP/ODP Hadoop monitoring capabilities on a tool called Pulse
Work closely with infrastructure, network, database and application teams at the client side
Monitoring and managing storage that is used by Hadoop system
Configuring security on Hadoop cluster
Worked with the multiple clients to install the Hadoop monitoring tool called Pulse, demonstrated the benefits of the Pulse and its daily usage
Technology Services Specialist
IBM
04.2017 - 02.2021
Managed and built several Hadoop clusters in production, development, Disaster Recovery environments
Involved in Hadoop CDH patching and CDH upgrade activities
Managed and built several Hadoop clusters in production, development, Disaster Recovery environments
Work with engineering software developers to investigate problems and make changes to the Hadoop environment and associated applications
Expertise in recommending hardware configuration for Hadoop cluster
Installing, Upgrading and Managing Hadoop Cluster on Cloudera distribution
Trouble shooting many cloud related issues such as Data Node down, Network failure and any OS related memory issues
Built and configured log data loading into HDFS using Flume
Helped developers in configuring Sqoop Importing and exporting data into HDFS and Hive tables
Recovering from node failures and troubleshooting common Hadoop cluster issues
Secured the cluster by enabling the Kerberos AD integration with cloudera manager and automated the TGT renewal and ticket generation for all the service accounts
Designed the users and groups creation based on the data loads and data tracks
Enabled the granular data access for those users by integrating sentry with the cloudera
Experience on decommissioning failed nodes and commissioning new nodes as the cluster grows and to accommodate more data on HDFS
Designed end to end backup disaster recovery for production cluster and make sure the replication should happened on time with the minimal data loss
Experience in enabling High Availability to avoid any data loss or cluster down time
Hands on experience in cluster upgrade and patching without any data loss and with proper backup plans
Managed and reviewed Hadoop Log files as a part of administration for troubleshooting purposes
Supporting Hadoop developers and assisting in optimization of mapreduce jobs, Sqoop-Hive Scripts, and HBase ingest, Kudu and Spark job failures if required
System/cluster configuration and health check-up
Continuous monitoring and managing the Hadoop cluster through Cloudera UI
Resolving tickets submitted by users, troubleshoot the error documenting, resolving the errors
Performed HDFS cluster support and maintenance tasks like Adding and Removing Nodes without any effect to running jobs and data
Experience in setting up high available to Hive Metastore, Hue
Hands on experience in managing and monitoring the cluster using Cloudera UI dashboard
Conducting root cause analysis and resolve production problems and data issues
Proactively involved in ongoing maintenance, support and improvements in Hadoop cluster
Document and manage failure/recovery (loss of name node, loss of data node, replacement of HW or node)
Involved in Minor and Major Release work activities
Executed tasks for upgrading cluster on the staging platform before doing it on production cluster
Installation and Administration of a Hadoop cluster
Experience in installing and configuring of HBase
Monitor Hadoop Name node Health status, number of Task trackers running, number of Data nodes running
Adding nodes into the clusters & removing nodes from the cluster for maintenance
Experienced in defining job flows
Experienced in managing and reviewing Hadoop log files
Environment: Hadoop, YARN, HDFS, HBASE, Flume, MapReduce, Sqoop, Zoo Keeper, Oozie, NoSQL, Oracle and Linux
DBA
Genpact India
06.2011 - 02.2015
Supported and assisted a team of DBA's in migration of databases from Oracle 9i to 10g to 11g
The role involved in installation, configuration & up-gradation of oracle databases using exp/imp utilities
Setup/configured oracle RMAN for database backup and recovery
Applied different recovery procedures for complete and incomplete recoveries
Cloned the instances from production to Test, Development instance to test & simulate few technical challenges, coordinate with testing and development teams
Performed Database Cloning and refresh on test and develop environment
Education
BSc -
Sri Venkateswara University
12th - Board of Intermediate Education
Narayana Junior College
10th - Board of Secondary Education
Margadarsi E M H School
Skills
Hadoop Architect
Certification
Certified AWS Solutions Architect Associate
Projectshandled
NucleusTeq, 05/2024 - Present, Participated in customer presentations as a Technical Architect to demonstrate capabilities in Big Data and AWS EMR Cloud technologies., Contributed to pre-sales efforts by preparing RfP responses, outlining technical solutions, and estimating resource requirements., Alation, 05/2023 - 02/2024, Expertise in designing, implementing, and managing Big Data solutions using technologies such as Hadoop, Spark, Kafka, Flink, etc., Hands-on experience with major Hadoop distributions (Cloudera, Hortonworks) and related ecosystem components.
Personaldossier
Currently seeking a challenging position where I can continue to uphold high performance standards.
Careerpath
NucleusTeq, 05/2024 - Present
Alation, 05/2023 - 02/2024
Acceldata Technology PVT LTD, 04/2021 - 05/2023
IBM, 04/2017 - 02/2021
TCS, 09/2015 - 02/2017
Future Focus PVT LTD. Deployed to client place TCS, 02/2015 - 09/2015