Summary
Overview
Work History
Education
Skills
Certification
Projects
Timeline
Generic

UMESH PANDURANG RATHOD

Mumbai

Summary

Accomplished Senior Manager at L & T Financial Services LTD, adept in leveraging Python and SQL for data transformation and analytics. Proven ability to optimize workflows and enhance scalability using GCP services. Strong problem-solving skills complemented by a collaborative approach, driving impactful results in data processing and integration.

Overview

9
9
years of professional experience
1
1
Certification

Work History

Senior Manager

L & T Financial services LTD
Mumbai
08.2024 - Current
  • Engineered scalable data transformation scripts using Python and Pandas to clean, enrich, and validate data before ingestion into BigQuery.
  • Wrote advanced SQL for analytical reporting, windowing functions, and optimizing queries for performance and cost-efficiency in GCP.
  • Built reusable Python modules for logging, alerting, and retry logic integrated with GCP services.
  • Deployed containerized Python services via Cloud Run for lightweight ML inference and REST APIs with autoscaling and zero-downtime.
  • Migrated legacy Flask apps to Cloud Run to enhance scalability and simplify CI/CD deployments using Cloud Build and Artifact Registry.
  • Developed and deployed streaming and batch data pipelines using Apache Beam (Python SDK) on Cloud Dataflow to process and transform real-time events from Pub/Sub to BigQuery, achieving sub-minute latency.

Senior Associate Technology

SYNECHRON TECHNOLOGIES PVT LTD
Pune
04.2024 - 06.2024
  • Leveraged Dataproc integration with other GCP services like Big Query and Cloud Storage to Optimize data ingestion, processing, and analysis workflows.
  • Strong proficiency in crafting SQL scripts tailored to specific requirements.
  • Create Python utilities used to connect source and target systems for data processing using PySpark.
  • Update existing PySpark code and attempt to make it more generic across the platform.

Consultant

CAPGEMINI TECHNOLOGY SERVICES INDIA LTD
Kuala Lumpur
12.2022 - 02.2024
  • Demonstrated proficiency in Google Cloud Platform components and Big Data Technologies.
  • Extensive experience in Spark code development for ETL processes and scheduler utilization.
  • Proficient in utilizing GCP services including Big Query, GCS, Airflow, Dataproc & Dataflow.
  • Integrated Dataflow pipelines with other GCP services such as Big Query and Cloud Storage to ensure seamless data ingestion and processing.

Officer Production

ETERNIS FINE CHEMICALS LTD
Pune
09.2021 - 02.2022
  • Utilized SQL to extract data from 8 different related tables from customer sales databases using JOIN and VIEW.
  • Developing and optimizing Oracle and MySQL queries to streamline data retrieval and analysis processes.
  • Troubleshooting and debugging SQL queries and performance issues.

Production Executive

VEDANT DYESTUFFS & INTERMEDIATES PVT LTD
Mumbai
06.2019 - 09.2021
  • Developed SQL queries to extract data from databases.
  • Created database schemas for newly generated data.
  • Responsible for debugging and troubleshooting data processing.

Associate

GEBBS HEALTHCARE SOLUTIONS PVT LTD
Mumbai
04.2016 - 03.2017
  • Developed SQL queries to extract data from databases.
  • Sort the extracted data in the Oracle database for further querying and analysis.
  • Proficient in utilizing GCP services including Big Query, GCS, Airflow, Dataproc & Dataflow.

Education

Bachelor of engineerING -

University of Mumbai
Mumbai, MH
05.2019

Skills

  • Data integration tools: Dataiku and Pajama
  • Big data technologies: Hive and BigQuery
  • Cloud services: Cloud Data Fusion, Cloud Composer, and Cloud Run
  • Google Cloud tools: Google Cloud SDK Shell, Google Cloud Functions, Google Cloud Dataflow, Google Cloud Pub/Sub, and Google Cloud Storage
  • Database management: Oracle, MySQL, and SQL Oracle
  • Scripting languages: Unix Shell Script and Python
  • Data processing frameworks: PySpark and Pandas
  • Version control: Git and GitHub
  • Project management: Jira and ADO

Certification

Associate Cloud Engineer (GCP), lle9f0, 105883

Projects

Nostradamus: Mumbai, MH, 08/01/24 - Present, 

Designed and optimized partitioned and clustered BigQuery datasets for petabyte-scale analytical workloads, reducing query costs by 40%

Developed ELT pipelines to ingest structured and semi-structured data into BigQuery using Dataflow and Cloud Storage.

Automated data ingestion and transformation tasks using Cloud Functions triggered by Cloud Storage and Pub/Sub, enabling near real-time processing.

Orchestrated custom ETL workflows on GCE VM instances with Python and Pandas, handling high-throughput batch data processing.

Used predefined Dataflow templates (e.g., GCS to BigQuery, Pub/Sub to BigQuery) to accelerate ETL pipeline deployment and integrated them with Cloud Composer for orchestration.

Containerized lightweight data services are deployed via Cloud Run for scalable, stateless microservices handling on-demand data transformation APIs 

Finance Cloud reporting platform: Pune, MH, 04/01/24 - 06/30/24, 

Troubleshooting and debugging Hive queries and performance issues

Hive partitioning, bucketing, and indexing for efficient data retrieval

Played a key role in data transformation and cleansing using Apache Spark operations

Designed and implemented a scalable data warehouse using BigQuery optimizing for performance and cost-efficiency Diagnosed and resolved issues by sampling data

Updated existing PySpark code to ensure compatibility with GCP Dataproc. 

Financial crime surveillance operation: Kuala Lumpur, MY, 12/23/22 - 02/29/24, 

Refactored on-premises Hive queries to be compatible with GCP BigQuery

Developed Python code for various Google Cloud API clients such as BigQuery, Configuration, Dataproc, and others Developed ETL pipelines to extract, transform, and load data from disparate sources

Created Python utilities to facilitate connections between source and target systems for data processing using PySpark

Migrated data from on-premises Hive to GCP BigQuery using the Dataiku tool

Converted Shell scripts into Airflow DAGs

Updated existing PySpark code to ensure compatibility with GCP Dataproc 

Customer sales analysis: Pune, MH, 09/01/21 - 02/28/22, 

Utilized SQL to extract data from eight different related tables from customer sales databases using JOIN and VIEW

Transformed and filtered data by using aggregating and filtering functions to improve the reporting process

Aggregated and visualized the data by using pandas to compile a professional report

Updated existing PySpark code to ensure compatibility with GCP Dataproc 

Data Quality Assurance: Mumbai, MH, 06/01/19 - 09/30/21, 

Implemented quality checks and automated processes to ensure the accuracy and completeness of healthcare data

Updated existing PySpark code to ensure compatibility with GCP Dataproc

Designed and maintained data warehouses to store and organize structured and unstructured healthcare data for analytics and reporting purposes

Developed Python code for various Google Cloud API clients such as BigQuery, Configuration, Dataproc, and others

Timeline

Senior Manager

L & T Financial services LTD
08.2024 - Current

Senior Associate Technology

SYNECHRON TECHNOLOGIES PVT LTD
04.2024 - 06.2024

Consultant

CAPGEMINI TECHNOLOGY SERVICES INDIA LTD
12.2022 - 02.2024

Officer Production

ETERNIS FINE CHEMICALS LTD
09.2021 - 02.2022

Production Executive

VEDANT DYESTUFFS & INTERMEDIATES PVT LTD
06.2019 - 09.2021

Associate

GEBBS HEALTHCARE SOLUTIONS PVT LTD
04.2016 - 03.2017

Bachelor of engineerING -

University of Mumbai
UMESH PANDURANG RATHOD