
Enthusiastic and eager to contribute to team success through hard work, attention to detail and excellent organizational skills. Clear understanding of requirements and tasks. Motivated to learn, grow and excel in Industry.
For application teams,
Design pipelines to load data from CosmosDB to SQL Server and vice versa for historical data, as well as data sync pipelines to sync data between different systems.
Implemented data transformation through Spark batch processing.
Developed and implemented performance improvement strategies and plans to promote continuous improvement.
Collaborated on ETL (Extract, Transform, Load) design, implementation, and maintenance of data integrity, and verifying pipeline stability.
Deriving analysis on various use cases for post-audit purposes in SQL, and projecting them in a Tableau dashboard.
Migrated on-prem SQL Server data warehouse to BigQuery.
For the Data Team.
Migrated legacy systems to modern big-data technologies, improving performance and scalability, while minimizing business disruption.
Streamlined complex workflows by breaking them down into manageable components for easier implementation and maintenance.
Designing and building a generic ELT pipeline, which can be reused for various data source ingestions to BigQuery using GCP and custom code.
Contributed to building a streaming pipeline with Kafka and Spark batch processing.
Conducted extensive troubleshooting to identify the root causes of issues and implement effective resolutions in a timely manner, using CI/CD orchestration tools like Airflow.
Agentic AI-
Reengineered existing ETL workflows to improve performance by identifying bottlenecks and optimizing code accordingly.
Currently, I am trying to build an automated version of the building pipeline based on our current implementation using Agentic AI.
The target is to reduce man-hours in creating a new pipeline and help them build ETL.
Cloud (GCP)