Summary
Overview
Work History
Education
Skills
Certification
Honorsandawards
Timeline
Generic

Rugved Vengurlekar

Summary

Highly skilled AWS Data Engineer with 5+ years of experience at Deloitte in designing and implementing Data Warehouse solutions, Cloud Migrations , Data Governance, Data Modelling for various clients across Banking, Insurance, Public Sector, Pharma and Telecom domain . Proficient in leveraging a range of Cloud Technologies including AWS Glue, AWS Step function, AWS Cloud Watch, AWS S3, AWS Redshift, AWS QuickSight, PySpark, Spark SQL , Python. ETL Tools :- AWS Glue. Database/ Data Warehouse :- Redshift, Oracle, SQL Server, My SQL, RDS.

Overview

5
5
years of professional experience
1
1
Certification

Work History

Data Engineer 2

JP Morgan Chase
05.2024 - Current
  • Designed and Architected Data Pipeline for reading Data from DynamoDb using Glue Job. Automated the pipeline using Event Bridge to generate events on daily basis and invoke lambda function to trigger glue job.
  • Generated processed parquet file in S3 bucket after transforming the data using aws glue. Created SNS topic, Lambda to trigger ECS task and loading the data Lake for Analytics.
  • Used Eac(Everything as code) / Terraform to create Infra.
  • Built Devops pipeline using Jenkins, Jules to integrate CI/CD
  • Used Control M to re-trigger lambda in case of failure.
  • Used Git for version control, Jira
  • Integrated Dynatrace, Splunk dashboard to monitor alerts and issues in production.
  • PySpark, Python, SparkSQL for script conversion.
  • Helped Wealth Engagement business to gain insights on Enrolled customers.
  • Connected with Multiple stakeholders to onboard new architecture to production.
  • Worked with product to build datapipline to provide insights for Personalizations customer business.
  • Used Boto3 to connect with DynamoDB, S3 etc.
  • Created Performance benchmarking for Glue, Dynamodb RCU, lambda, s3 cost and runtime with difference Glue processor, Dynamodb RCU configurations with saved AWS services cost by 70% and decreased the datapipeline runtime.
  • Created Global secondary index on Dynamodb to read 70 million records using batch_get_item and query data. Worked on multiple approaches to scan data using multiple segment scan.

Consultant - Senior Data Engineer

Deloitte USI
05.2022 - 04.2024
  • Designed and developed scalable data pipelines and ETL processes using AWS data integration and transformation services, ensuring efficient and reliable data processing
  • Design, Development and Deployment of core platform components
  • Building and implementation of ETL pipeline for client using AWS Glue, Glue Crawler, Lake Formation, Data Catalogue, S3, Athena, DynamoDb
  • Worked on generating Client reports using Power BI for data analytics
  • Created Data Model for Multi Model dataset by organizing and structuring data from different sources and formats into a unified model that can be used for analysis or reporting
  • Worked and guided resources on data migration project for migrating data from Netezza to Redshift for Banking client
  • Worked on AWS Glue, S3, Redshift, Airflow and wrote ETL PySpark scripts to perform transformation and migrate data to Redshift
  • Designed, implemented and automated Audit mechanism process for data loads which reduced time & effort for data validation by 95%
  • Also worked on Spark memory optimization which helped in reducing the glue job run time and job efficiency, hence saving operational cost
  • Collaborated with cross-functional teams, including developers, testers, and stakeholders, to gather requirements, design solutions, and ensure timely project delivery
  • Used BitBucket, Github for code version control and performed code reviews
  • Participated in code reviews, providing technical guidance and feedback to team members, and mentoring junior team members when needed
  • Used Jira for issues and project task tracking

AWS - Data Engineer

Principal Financial Group
09.2021 - 05.2022
  • Worked on AWS Glue, Athena, S3, Cloudwatch, Stepfunctions, Cloudformation, Azure devops
  • Worked on building data pipeline to generate client reports for Insurance domain
  • Have worked on integrating data from multiple sources and transforming data to required format to generate client reports using Tableau
  • Worked with onshore counterpart and clients to get the requirements and implemented the required Data transformations to get the data as required for generating reports
  • Built Data Pipeline using Stepfunctions, S3, Glue, lambda, CDK, Cloudformation CloudWatch, Secret Manager etc to generate client reports
  • Worked on resolving memory issues in spark dataframe on glue caused due too huge size of data

Application Development Analyst

Accenture
08.2019 - 09.2021
  • Leveraged popular Python libraries, including Pandas to manipulate, analyze, and perform scientific computing tasks on large datasets
  • Integrated with cloud services and APIs, such as AWS Boto3, to interact with cloud resources and services programmatically, enabling efficient and automated workflows within cloud-based environments
  • Written PySpark script to perform data transformation
  • Worked on developing mapping using Informatica Powercenter to load data to AWS Redshift.
  • Worked on designing the end-to-end ETL pipeline using AWS glue, Step-Functions, S3, Dynamo Db, RDS, Lambdas etc
  • Contributed fully to design assessments and code reviews
  • Written Unix Shell Scripts to Migrate all the on-premises archive data files to AWS Glacier

Education

Master of Technology - Software Engineering

Birla Institute of Technology and Science
06.2023

Bachelor of Engineering - Computer Engineering

Sinhgad College of Engineering
01.2019

Skills

  • Data Migration
  • Pyspark
  • Amazon Web Services (AWS)
  • Database Management System
  • AWS Glue
  • Solution Architecture
  • AWS Step Functions
  • Amazon Redshift
  • Python (Programming Language)
  • SQL

Certification

  • AWS Certified Developer - Associate, Amazon Web Services (AWS)
  • Microsoft Certified: Azure Data Fundamentals, Microsoft
  • Microsoft Certified: Azure AI Fundamentals, Microsoft
  • Microsoft Certified: Azure Fundamentals, Microsoft

Honorsandawards

  • Applause Award, Deloitte, 08/01/23
  • Zenith Award, Accenture

Timeline

Data Engineer 2

JP Morgan Chase
05.2024 - Current

Consultant - Senior Data Engineer

Deloitte USI
05.2022 - 04.2024

AWS - Data Engineer

Principal Financial Group
09.2021 - 05.2022

Application Development Analyst

Accenture
08.2019 - 09.2021

Bachelor of Engineering - Computer Engineering

Sinhgad College of Engineering

Master of Technology - Software Engineering

Birla Institute of Technology and Science
Rugved Vengurlekar