Summary
Overview
Work History
Education
Skills
Projects
Certification
Achievements And Awards
Languages
Timeline
Generic

Rahul S Desai

Pune

Summary

Hardworking employee with customer service, multitasking and time management abilities. Devoted to giving every customer a positive and memorable experience. Diligent Senior Bigdata Developer with 8.6 years of experience in bigdata stack namely Hadoop, Spark and SQL.

Overview

7
7
years of professional experience
1
1
Certification

Work History

Senior Associate Technology

Synechron Technologies Pvt Ltd
Pune
09.2021 - Current
  • Total 6.9 years of exclusive experience in Hadoop and spark as well as its components like Hdfs, Hive, Sqoop, Spark-Core, Spark-SQL. Also good exposure in SQL, Jenkins CI pipelines using bitbucket and git bash
  • Developed and implemented software solutions to improve operational efficiency.
  • Analyzed and troubleshot technical issues, identified root cause of problems, and provided resolutions.
  • Good exposures in Impala, Starburst, Autosys and Control-M Scheduler ,Urban Deploy, Agile scrum methodology and jira
  • Good exposure in Banking(Institutional) domain .Excellent communication, interpersonal, analytical skills, and strong ability to perform as part of team
  • Exceptional ability to learn new concepts, Hard working and enthusiastic
  • Have been rewarded with one of the most rated Award in the Organization i.e SURPASS Award for the outstanding performance.

Hadoop and Spark Developer

Legato Health Technology
Hyderabad
11.2019 - 09.2021
  • 1.8 years of exclusive experience in Hadoop and Spark as well as its components like HDFS, Hive, Sqoop, Spark-Core, Spark-Sql
  • Exclusive experience in data transformation from on premise (Hadoop/hive) to cloud (AWS and Snowflake) using hummer and Vulcan framework
  • Knowledgeable in Pig, Flume, Kafka, Spark-Streaming, Agile scrum methodology and jira.
  • Designed and developed data pipelines to ingest structured data into HDFS using Apache Spark.
  • Tuned Spark configurations to achieve optimal execution times of batch processes.
  • Created Hive tables, optimized queries, stored procedures, functions and views on Hadoop clusters.
  • Developed and implemented Spark applications using Scala.

Software Engineer

Tritech Software Pvt Ltd
Hyderabad
07.2017 - 11.2019
  • 2.5 years of exclusive experience in Hadoop and Spark as well as its components like HDFS, Hive, Sqoop, Spark-Core, Spark-Sql
  • Import all data from RDBMS to HDFS for further processing
  • Writing the script files for processing data and loading to HDFS
  • Created Hive tables to store the processed results in a tabular format
  • Batch processing of data sources using Spark
  • Implemented Spark RDD transformations, and actions to implement business analysis
  • Implemented Spark using Scala and Spark SQL for faster testing and processing of data
  • Implemented partitioning, dynamic partitions and buckets in HIVE
  • Involved in gathering the requirements, designing, developing and testing.

Education

B.Tech/B.E. -

Savitribai Phule Pune University
Pune
01.2017

12th -

01.2013

10th -

01.2011

Skills

  • Hadoop
  • Spark
  • HDFS
  • Hive
  • Spark Core
  • Spark SQL
  • Impala
  • Starburst SQL
  • Sqoop
  • Kafka
  • AWS
  • UNIX Shell Scripting
  • Jenkins CI
  • Bitbucket
  • Urban Deploy
  • AutoSys Scheduler
  • Control-M
  • Adaptability and Flexibility
  • Problem-solving abilities

Projects

#Project 1: APAC Reporting - Sybase Migration

APAC Reporting is a migrated application where all the data and logic has been migrated from Legacy (Sybase DB) to Olympus as centralized location. Olympus is getting trading data from different source systems. As part of the same business each client would be offered with different types of modules based on their needs. To maintain this much of huge volumes of, different verities of data in traditional databases is a very tedious process. Re-platforming of current data warehouse system to Hadoop solution in a cost effective solution. Generating different types of regulatory reports which can be helped client to process their analysis w.r.t of the regulators. 

#Project 2: CII-Client Information Insights

 CII is getting the source data from different source systems. As part of the same business each customer would be offered with different types of modules based on their needs. To maintain this much of huge volumes of, different verities of data in traditional databases is a very tedious process. To meet the scaling needs of data of CII, re-plat forming of current data warehouse system to hadoop solution in a cost effective solution. Currently, Snowflake has been implemented to get the data for visualization (Teradata system has been replaced with Snowflake). 

#Project 3: JSK-Data Analysis

The purpose of the project is logistics providers are persistently creating enormous and vast data sets while managing the massive flow of goods and individuals. For millions of shipments around the world every day, location, content, size, weight, origin and destination and many other information are getting tracked and gathered across the global delivery and transport networks. Maintaining this much of the volume of different verities of the data in traditional databases is a very tedious process. To meet the scaling need of data of jsk logistics, replatforming the current data warehouse system.

Certification

  • AWS Certified Cloud Practitioner
  • AWS Certified Developer - Associate

Achievements And Awards

  • SURPASS Award - Q4 2023
  • STAR Award - Jan'22
  • Outstanding Student for year 2017

Languages

  • English
  • Hindi
  • Marathi

Timeline

Senior Associate Technology

Synechron Technologies Pvt Ltd
09.2021 - Current

Hadoop and Spark Developer

Legato Health Technology
11.2019 - 09.2021

Software Engineer

Tritech Software Pvt Ltd
07.2017 - 11.2019

B.Tech/B.E. -

Savitribai Phule Pune University

12th -

10th -

Rahul S Desai