Dynamic Senior Manager with a proven track record at L & T Financial Services, specializing in data pipeline development and cloud architecture. Expert in leveraging Vertex AI and GitHub Copilot to enhance productivity and automate workflows. Adept at mentoring teams, driving innovation, and ensuring data quality through advanced transformation techniques.
Overview
5
5
years of professional experience
Work History
Senior Manager
L & T Financial Services Pvt Ltd
Mumbai
04.2024 - Current
Automated drafting of MRM (Model Risk Management) and BRD (Business Requirements Document) using GitHub Copilot, accelerating documentation workflows by 40%.
· Leveraged GitHub Copilot to accelerate development of ML pipeline components, reducing coding time and improving productivity by ~30%
· Developed and deployed end-to-end Kubeflow Pipelines on Vertex AI to automate model training, evaluation, and deployment workflows.
Design and execute complex data transformation and cleansing pipelines using Apache Spark, ensuring data quality and consistency.
Enhanced and refactored existing PySpark scripts to improve performance and ensure full compatibility with GCP Dataproc.
Design and execute complex data transformation and cleansing pipelines using Apache Spark, ensuring data quality and consistency.
Developed and deployed end-to-end Kubeflow Pipelines on Vertex AI to automate model training, evaluation, and deployment workflows.
Containerized Python-based ETL and ML inference applications using Docker, enabling consistent and portable deployments across environments.
Optimized container startup and execution time to improve performance and cost efficiency of jobs deployed on Cloud Run.
Senior Associate Engineer
Synechron Pvt Ltd.
Pune
10.2021 - 01.2024
Company Overview: Synechron offers IT consulting and business process outsourcing services to various industries, such as banking, healthcare, manufacturing, media, and entertainment. It is globally recognized as a leader in data warehouse modernization and migration.
Migrated Teradata database to BigQuery using GCP tools such as BigQuery, GCP Composer, DataFusion (Pipelines) and Google Cloud SDK Shell, as well as Cloud Function.
Created and Datafusion (Pipeline) for exporting and importing data into GCP (BigQuery) and resolved performance issues.
Scheduled Ingestion Dags and using Cron Scheduler.
Developed BigQuery authorized views for row-level security and data exposure to other teams.
Utilized Temporary Tables to cache query results not written to permanent tables.
Employed Pub/Sub and Dataflow to load streaming and real-time data from various sources to BigQuery.
Established multiple cloud storage buckets in GCS for file storage.
Implemented Cloud Functions to trigger file loading from CSV files to BigQuery at any time.
Crafted complex SQL queries using Joins and optimized query performance through clustering and partitioning.
Mentored and guided newcomers on project-related matters.
Conducted Unit testing to validate complex code and deliver it to clients.
Utilized Google Cloud SDK Shell to transfer data between environments.
Leveraged Unix commands to navigate through script and KSH file logs.
Utilized Control-M for scheduling jobs, determining timing, frequency, and dependencies.
Utilized Git (Bitbucket) Version Control Software for deploying to Prod Environment, creating master and feature branches.
Employed Terraform for Hot Fixing and raised PRs for deploying Table and view definitions, as well as JSON files of Data Fusion (Pipeline) and dags.
Transitioned code development from On-Prem technologies like Datastage and Informatica to BigQuery platform.
Orchestrated and scheduled DAGs using Cloud Composer.
Conducted reviews of deliverables to ensure adherence to quality standards.
Maintained comprehensive project documentation, including project-specific details and mapping documents.
Automated data loading from Google Cloud Storage (GCS) to BigQuery using Dataflow for efficient and scalable processing.
Employed Dataflow to compress large files in bulk before loading into BigQuery, reducing storage costs and improving data transfer efficiency.
Implemented tokenization and redaction of sensitive data using Data Loss Prevention (DLP) API template-based data masking to safeguard Personally Identifiable Information (PII) and ensure compliance with data regulations.
Utilized Pub/Sub as a messaging service to export Google Cloud Platform (GCP) logs to BigQuery via a log router, maintaining a comprehensive history of specific batches or events for analysis and monitoring.
Established Database Replication with Change Data Capture (CDC) to replicate SQL data into BigQuery, ensuring real-time updates and synchronization between source and destination data stores.
Leveraged Data Fusion and ArgSetter to build reusable data integration pipelines, enabling parameterization and adaptability across diverse environments and scenarios.
Designed and implemented a data pipeline using Data Transfer Service to seamlessly load data from SQL databases into BigQuery, streamlining data migration and synchronization processes.
Synechron offers IT consulting and business process outsourcing services to various industries, such as banking, healthcare, manufacturing, media, and entertainment. It is globally recognized as a leader in data warehouse modernization and migration.
Project Coordinator
ManpowerGroLlp Services India Pvt Ltd
Pune
02.2020 - 08.2021
Company Overview: Jota Aviation, the specialist charter and cargo airline.
Developed views within BigQuery tables to streamline data analysis and reporting processes.
Implemented batch data processing methods for efficient handling of large datasets.
Played a key role in resolving hotfix issues to maintain data pipeline stability and performance.
Designed and executed cloud-to-cloud data workflows, enabling seamless data retrieval across different regions.
Engineered intricate workflows for extracting and transforming data between BigQuery instances, optimizing data analytics capabilities.
Actively participated in code reviews to ensure ETL and database components met global standards and compliance requirements.
Ensured proper exception handling with detailed error descriptions and error locations in all stored procedures.
Transferred data from GCP to the mainframe in binary file format or tape format using scheduled JSON and export JSON.
Scheduled division-wise Control M jobs in Control M and incorporated proper module dependencies within the same Control M jobs.
Prepared unit test documents for stored procedures and documented over 20 test cases in unit testing artifacts.
Created unit test documents for scheduled Control M jobs, including job execution timing and frequency.
Deployed code to production using CICD Jenkins.
Reviewed deliverables to ensure compliance with quality standards.
Refined and complex SQL queries were developed utilizing joins, subqueries, and correlated subqueries to efficiently retrieve data from the database.
Created Oracle Stored Procedures and Functions to seamlessly integrate business rules.
Engaged in scripting SQL according to client specifications.
Worked with Implicit Cursors, Explicit Cursors, and Ref Cursors.
Managed tables through Views, Sequences, and Indexes.
Facilitated the importation of data from client files into Oracle.
Participated in implementing new changes and defect resolution.
Conducted unit testing as part of the development process.
Jota Aviation, the specialist charter and cargo airline.
Manager-Regulatory Reporting & Bank Reconciliation at AXIO (brand name of CapFloat Financial Services Pvt. Ltd)Manager-Regulatory Reporting & Bank Reconciliation at AXIO (brand name of CapFloat Financial Services Pvt. Ltd)
CASHIER CUM ACCOUNTS at ORIENT EXCHANGE AND FINANCIAL SERVICES PVT LTD - THRISSURCASHIER CUM ACCOUNTS at ORIENT EXCHANGE AND FINANCIAL SERVICES PVT LTD - THRISSUR