Summary
Overview
Work History
Education
Skills
Certification
Languages
Hobbies and Interests
Projects
Personal Information
Disclaimer
Timeline
Generic

Mrunal Patil

Pune

Summary

Dedicated Data Engineer with 3+ years of expertise in designing and implementing robust data solutions using AWS cloud services and big data technologies. Proven track record of developing scalable data pipelines, performing complex data migrations, and optimizing data processing workflows. Demonstrated ability to transform complex data challenges into efficient, automated solutions using Python, PySpark, and AWS services.

Overview

4
4
years of professional experience
1
1
Certification

Work History

Data Engineer

TATA Consultancy Services
03.2021 - Current
  • Worked across multiple projects involving the design and implementation of data solutions using AWS and big data technologies.
  • Played a key role in all phases of the projects, from data migration to optimizing data workflows, ensuring the delivery of high-quality, scalable solutions.
  • Built a model monitoring solution to track data drift, performance, and explainability using databricks, Pandas, SHAP and Azure enabling continuous monitoring and improved governance.

Education

Bachelor of Engineering - Electronics And Communications Engineering

Jain College of Engineering,Belgaum
Belgaum,Karnataka
09-2020

Skills

  • Python
  • SQL
  • Pyspark
  • AWS
  • Azure
  • Data Factory
  • Azure Synapse Analytics
  • Databricks
  • Data Lake
  • S3 Bucket
  • Glue
  • Redshift
  • Lambda
  • KMS
  • Cloudwatch
  • EMR
  • Kinesis
  • GitHub
  • MySQL Workbench
  • Visual Studio Code
  • Jira
  • PyCharm
  • Machine Learning

Certification

  • Azure Data Engineer certification (DP-203) (Training)
  • AWS Certified Data Engineer – Associate, DEA-C01
  • Python and Spark training, E2 competency
  • Databricks Certified Data Analyst Associate

Languages

  • English
  • Hindi
  • Marathi
  • Kannada

Hobbies and Interests

  • Cooking
  • Traveling
  • Drawing

Projects

1. Data retention and destruction

  • Role: big data developer
  • Technology: big data
  • Project summary

             Implemented a data deletion strategy for securely managing data and case files over 10 years old This initiative supports annual data purges, ensuring compliance with data retention policies and optimizing storage.

2. L1 migration

  • Role: Developer
  • Technology & tools: Python, Pyspark, Redshift, Glue, SQL, JIRA, VS Code, GitHub
  • Project summary

         The project involved migrating data from a Netezza database to Amazon Redshift using AWS Glue. Our main responsibility was to ensure the data was compatible with Redshift's structure. To do this, we created a configuration file that defined how data should be transferred, validated it for accuracy, and provided this as an input to the execution team. This configuration helped automate the migration process, ensuring smooth data movement and accuracy. Once validated, we used Glue to load the data into Redshift, completing the migration efficiently and securely.

3. Data Services Framework

  • Role: Developer
  • Technology & tools: Python, Pyspark, Glue, Redshift
  • Project summary

       Our project involved building a data-loading framework to move data between different layers (L1, L2, and L3) within our data pipeline. The goal was to create a flexible, reusable framework that could handle any type of incoming data. This required designing processes that automate data validation, transformation, and loading across all layers, ensuring compatibility and integrity regardless of data type. By developing this robust framework, we enabled consistent, automated data handling, which saves time and reduces errors in the data pipeline.

4. AI/ML model monitoring and explainability framework

  • Role: Data Engineer / AI/ML Engineer
  • Technology and tools
  • Python, Databricks, PySpark, SHAP, Population Stability Index (PSI), ML Monitoring Frameworks, and Cloud Reporting
  • Project summary

Developed an AI/ML model monitoring framework to track data drift, feature stability, and model performance in production. Implemented PSI, feature drift checks, and SHAP explainability to identify key drivers influencing predictions and ensure transparency. Automated monitoring and reporting through Databricks pipelines, enabling continuous tracking, improved governance, and early detection of model drift

Personal Information

  • Date of birth: 13 July 1996
  • Gender: Female
  • Address: A1-902 Aishwaryam Courtyard Phase-2, Newale Vasti, Chikhali, Pune

Disclaimer

I hereby declare that the above written particulars are true and correct to the best of my knowledge and belief.

Timeline

Data Engineer

TATA Consultancy Services
03.2021 - Current

Bachelor of Engineering - Electronics And Communications Engineering

Jain College of Engineering,Belgaum
Mrunal Patil