Summary
Overview
Work History
Education
Skills
Timeline
Generic
GAJENDRA SINGH

GAJENDRA SINGH

Gurugram

Summary

Lead Data Engineer and Data Architect with over 11 years of experience, including 6+ years in Big Data technologies. Demonstrated expertise in delivering data-driven solutions that improve efficiency, accuracy, and scalability of data processes. Proficient in designing and implementing end-to-end data pipelines, performing data analysis, and developing visualizations and reports to transform complex data into actionable business insights.

Overview

11
11
years of professional experience

Work History

Lead Data Engineer

Tredence Inc
Gurugram
09.2023 - 08.2025
  • Developed a data pipeline for a multinational pharmaceutical company in Azure Databricks using the Medallion architecture to efficiently transfer asset, batch, and quality data from the source, AWS S3, to Azure ADLS Gen2.
  • Data ingestion was performed using Databricks Workflows and Azure Data Factory.
  • Implemented all data transformations using PySpark and Spark SQL. Leveraged Databricks Autoloader to handle streaming jobs, enabling real-time data ingestion from ADLS Gen2.
  • Redesigned and enhanced the CI/CD process using Databricks Asset Bundle and Azure DevOps, ensuring streamlined deployment and automation.
  • Set up a Databricks Genie workspace to facilitate data analysis in a conversational manner.
  • Parameterized ingestion and transformation scripts to seamlessly accommodate new tables and columns.
  • Ensured proper use of Unity Catalog for robust data governance, enabling centralized access control, auditing, and lineage tracking across the data pipeline. Implemented fine-grained permissions and data classification to enhance security and compliance.
  • Led a team of 10 professionals, providing technical guidance, mentorship, and strategic direction to ensure the successful execution of projects

Senior Solution Advisor

Deloitte USI
Gurgugram
10.2021 - 09.2023
  • Designing and building robust solutions for storing very large amount of raw data of different sizes using Blob Storage and ADLS
  • Building Spark Applications using PySpark and Spark-SQL in Databricks for data Extraction, Transformation and Aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns
  • Design and Implement data pipeline to process structured and semi-structured data from different sources using ADF and Databricks
  • Proficient with scripting in PySpark/Python to extract data from various formatted files like Parquet, JSON, CSV etc
  • Work on writing complex queries and creating views on PostgreSQL for data visualization team
  • Manipulating table structure in PostgreSQL as per latest needs and adjusting the existing data as demanded
  • Designed complex QC process to validate integrity and accuracy of extracted data using Python and PostgreSQL

Assistant Manager

Bank of America
Gurugram
09.2019 - 10.2021
  • Developed robust pipeline with the help of Azure Databricks/PySpark, Data Factory and Azure Data Lake for monthly/weekly balances and flows at product, division and market level
  • Data ingestion to one or more Azure Services like Azure Data Lake, Blob Storage, Azure SQL and Azure DW and processing the data by using Azure Databricks
  • Building and designing data models and solution design for Business requirements
  • To meet specific business requirement wrote UDF in PySpark
  • As part of business requirement flows and balances reporting at various investment solution level is done for CFO group with the help of Python, SAS and SQL
  • Market level Variance analysis of product balance on quarter is done and then data is made available to local market leadership, market support personnel and public sector bankers

Business Analyst

Conduent
Gurugram
10.2018 - 09.2019
  • Writing SAS/SQL program to fetch required data from Database tables as per requirements
  • Reading the raw data using SAS Procs like PROC IMPORT
  • Used SAS for pre-processing of data, data extraction, validation & manipulation
  • Created temporary and permanent SAS datasets using temporary and permanent Libraries
  • Used SAS/Base for reporting purpose and SAS/MACROS for automating the same
  • Designing SAS SQL queries to generate report to share with the Client Team
  • Developed Tableau dashboards for better visualization of data using charts and graphs and enable effective business decisions

Configuration Analyst II Exp

Aon
Gurugram
03.2014 - 10.2018
  • Creating the SAS programs to generate the daily reports for different clients
  • Verification of the fixed issues identified in the testing phase
  • Provide support for the deployed project till it is stabilized
  • Involved in reviewing the test cases and executing the same
  • Designing SQL queries while performing database testing
  • Wrote complex SQL queries using joins, sub queries and correlated sub queries

Education

Master of Computer Applications -

College of Engineering Roorkee
Roorkee, Uttrakhand
07.2013

Skills

  • Azure Databricks
  • Azure Data Factory
  • Azure Data Lake Storage
  • Azure SQL
  • Azure DevOps
  • Azure Key Vault
  • PySpark
  • Python
  • SQL
  • GitHub
  • Solution design
  • Data governance

Timeline

Lead Data Engineer

Tredence Inc
09.2023 - 08.2025

Senior Solution Advisor

Deloitte USI
10.2021 - 09.2023

Assistant Manager

Bank of America
09.2019 - 10.2021

Business Analyst

Conduent
10.2018 - 09.2019

Configuration Analyst II Exp

Aon
03.2014 - 10.2018

Master of Computer Applications -

College of Engineering Roorkee
GAJENDRA SINGH