Summary
Overview
Work History
Education
Skills
Projects
Certification
Personal Information
Timeline
Generic
Rajasankar Radhakrishnan

Rajasankar Radhakrishnan

Data Architect/Data Engineer
Bengaluru

Summary

A seasoned Data Architect and Data Engineer with over 12 years of experience designing and implementing data platforms on hyperscalers like Azure and GCP. Skilled in utilizing cloud-native and open-source technologies to support leading automotive, manufacturing, and BFSI clients. Proficient in modern data architectures such as Lakehouse, DataMesh, and active metadata management, with a strong focus on data quality, automation, and integrity. Adept at designing metadata-driven solutions based on the Medallion architecture with integrated data governance. Additionally, possesses extensive knowledge of traditional dimensional modeling and data warehousing, along with expertise in Python/SQL-based analytics solutions. A proven leader, having successfully led cross-functional teams, driven digital transformation, and defined data roadmaps to deliver strategic value for clients.

Overview

13
13
years of professional experience
3
3
years of post-secondary education
3
3
Certifications

Work History

Data architect

BGSW(Robert Software Global Software )
01.2021 - Current

Key Technologies worked:

Data architecture,DataBricks,DataFactory,SnowFlake, Azure, GCP, Data Lake, Delta Lakehouse, Data Governance,Python,NoSql

Key Responsibilities:

  • Spearheaded cloud migration initiatives for Greenfield and Brownfield Manufacturing Plants across the globe managing vast amounts of enterprise data assets.
  • Designed and Created Enterprise Data Lake, Delta Lakehouse with PySpark, Spark SQL for processing large volumes of data.
  • Supported in Business Acquisition by analyzing RFPs,RFQs and providing Proposals for potential customers.
  • Led Data Engineering Tech CoE actively taking part in numerous PoCs.


Data Engineering Specialist

BGSW(Robert Software Global Software )
01.2018 - 12.2021

Key Technologies worked:

Azure Databricks,Azure DataFactory,Azure IOTHUB,Azure Event hub,Grafana,Stream analytics,PowerBI

Key Responsibilities:

  • Enhanced data quality by performing thorough cleaning, validation, and transformation tasks.
  • Automated routine tasks using Python scripts, increasing team productivity and reducing manual errors.
  • Fine-tuned query performance and optimized database structures for faster, more accurate data retrieval and reporting.
  • Experienced in Design and development of analytic applications using programming languages such as Python, Pyspark, SQL, KQL etc.
  • Experienced in developing data pipelines using ETL tools/services such as SSIS, Azure data factory, Azure data bricks, Azure Synapse.
  • Experienced in writing Spark RDD transformations, actions for processing data from sources such as IOT hub, Event hub, data lake gen2, SQL databases, delta, parquet and json files etc.

Data Engineer

BGSW(Robert Software Global Software )
06.2016 - 12.2017

Key Technologies worked:multi collinearity

ADLS Gen2, Azure SQL, IOT hub, Azure ML,ADB,SSIS,PowerBI,Grafana

Key Responsibilities:

  • Azure Data engineering/ML/ETL/Visualization/Application development /Database programming.
  • Implemented API integration for seamless communication between different software components.
  • Performed descriptive analytics using numpy, pandas, scikit-learn in Python to understand relationship between variables and to detect outliers, Interaction and multicollinearity among different dimensions of raw materials.
  • Experience working with Other Azure services such as ADLS Gen2, Azure SQL, IOT hub, Azure ML,ADB,Synapse
  • Experienced with Reporting visualization tools such as Power BI, Grafana. crystal Reports etc.

Assistant Manager-IT(ETL Developer)

Titan company limited
4 2015 - 6 2016

Key Technologies worked:

SSIS,SQL,SSRS,Crystal Reports

Key Responsibilities:

  • Designed and developed Data warehouse for Titan Management information system.
  • Upgraded existing Crystal reports to SSRS and Power BI reports resulting in 75% decreased runtime of reports.
  • Used SSIS SCD components to capture slowly changing dimensions to generate historically accurate reports.
  • Upgraded existing C# /HTML reports by Converting inefficient Microsoft SQL Queries to Recursive CTEs and efficient Stored procedures.
  • Used Snapshot and Cache mechanism of SSRS to reduce load on data warehouse.

Executive-IT(SQL Developer)

Titan company limited
08.2012 - 03.2015

Key Technologies worked:

SQL Development,Performance Optimization,Application development

Key Responsibilities:

  • Analyzed existing SQL queries to identify opportunities for improvements.

Executive Trainee-IT

Titan company limited
07.2011 - 07.2012
  • Application development/ Database programming.

Education

B.E.(computer science) - Computer Science

Vel MultiTech Engineering College(Anna University
Chennai
06.2008 - 05.2011

Skills

SQL

ETL development

Data Analytics

Data Modeling

Real-time Analytics

Data integration

Data Lakehouse

Projects

1.Bosch Semantic Stack Data Layer(Oct 2023- till date):

Role: Data Architect/Senior Data Engineer

Project Description: Unified Data Platform Connective Automotive use cases in Bosch.


  • Defined, Architected and Implemented a self-serve multi-tenant data platform to Fasttrack advanced analytics for automotive and manufacturing use cases.
  • Analyzed multiple options(Fabric, Snowflake, Databricks) and shortlisted the best tools and technology based on different parameters in collaboration with product owners (PO) and management, ensuring alignment with business objectives.
  • Led the development with two PODs focusing on delivery of Features such as Configurable Data pipeline Orchestration Framework, Delta Lakehouse Optimization, COVESA Standardized semantic layer etc.
  • Implemented a user-friendly Framework for Role based access control, Row level and column level security using Databricks Unity Catalog on the data Layer.
  • Provided technical mentorship, fostering team development through code reviews and guidance on best practices across API development Life cycle, Data archival framework.
  • Collaborated with cross-functional teams to design and implement data extraction, transformation, and loading (ETL) processes in alignment with business objectives.

Tech stack: Azure, Databricks, ADF,ADLS Gen2,Azure Devops




2.CIMB-DataCore(Jan 2023-Nov 2023)

Role: Data Architect/Senior Data Engineer

Project Description: Real Time Lakehouse for Gold standard Master data consumption


  • Setting up Unity catalog and the different data layers of the Lakehouse.
  • Design and Implementation of real time data pipelines for data ingestion from kafka till consumption layer using Delta Live tables for low latency data availability.
  • Implementation of different Data governance capabilities such as Role based access control, RLS and CLS using Databricks dynamic views and SQL functions.
  • Implementation of configuration driven DLT module for easier onboarding on new data pipelines.


Tech stack: GCP Databricks, Unity catalog, GCS, Confluent Kafka, Spark   structured streaming.


3.VHIT-Lake house implementation(July 2022 – April 2023)

Role: Data Architect/Senior Data Engineer

Project Description: Real Time Lakehouse for Gold standard Master data consumption

  • Architecting the solution for Data-lake implementation
  • Creation of Technical design document, Test case document, user manual etc.
  • Designing and implementing the dynamic data pipelines using ADB, ADF, SQL server, ADLS gen2 for validating the files, checking the data quality for completely metadata driven data lake.
  • Creation of mount points, authenticating access between Azure resource using Service principle, managed identities, and key vault.
  • Analyzing and optimizing the data loads for larger tables and files by tuning spark configuration, applying zorder, optimize, vacuum, partitioning.
  • Monitoring spark jobs using Ganglia UI, Spark UI to understand the bottlenecks and optimize the jobs.
  • Creation of Pyspark notebooks for validating files, data and business logic transformations and creation of wheel files for reusable code.
  • Creation of ADF pipelines for orchestrating the various


Tech stack: Azure Databricks, Azure data factory

4.DFL-2.1(Nov 2021– June 2022)

Role: Data Architect/Senior Data Engineer

Project Description: Migration of existing Big data application from Synapse to Databricks and implementation of Data lake for processing data from 5 different applications.


  • Conversion of existing Synapse stored procedures as Pyspark and Spark SQL notebooks.
  • Reduction of data load time by 75 % and cost savings by over 80% compared to the existing Data Warehouse.
  • Implementation of custom framework to do incremental load from Source databases even in the absence of watermark columns in source systems.
  • Smart spark job parallelization framework based on source table row count to reduce load time.
  • Implemented Json read module to dynamically identify and decode encrypted json files from Iothub.
  • Implemented Lifecycle management for Datalake to set appropriate retention periods and layers (hot, cold, archive) to reduce storage cost.
  • Creation of Grafana monitoring reports using Azure log analytics as source and KQL for transforming data.

Tech Stack: Azure Databricks , Azure data factory, ADLS gen2,Azure SQL server, Pyspark , IOTHUB, PowerBI , Grafana




5.DataNexus: Factory of Future(Feb 2020– June 2021 )

Role: Data Engineer Specialist

Project Description: Migration of existing Big data application from Synapse to Databricks and implementation of Data lake for processing data from 5 different applications.


  • Implementation of Authentication using which multiple services can access each other securely using managed identities, service principle and key vault.
  • Creation of File validator module using higher order functions which lets users configure dynamic File validation rules.
  • Creation of Data quality validator module which lets users configure dynamic Data quality validation rules.
  • Processing IOT json files using Databricks Autoloader(cloudfiles) for maintaining a near real time gold layer.
  • Creation of Mock data generation framework for removing the data non-availability hindrance.
  • Creation of Alert framework to inform stakeholders of job failures/validation failures as per alert configuration.

Tech Stack:

Azure Databricks , Azure data factory, ADLS gen2,Azure SQL server, Azure functions. Pyspark


Mixing 2.0(Jan 2018 – Aug 2020)

Role: Data Engineer

Project Description: Implementation of recipe generation algorithm to shortlist best set of raw materials (from half a million possibilities) from stock warehouse for Push belt production. The objective to is to enhance the pushbelt quality and reduce scrap.

  • Created a custom algorithm(using numpy , pandas and statsmodels library) in Python to shortlist best set of raw materials(from half a million possibilities) from stock warehouse for Push belt production, resulting in time reduction of stock preparation stage by 75%.
  • Created and deployed a suite of microservices in python for cleaning stock data, predicting missing measurements and identifying best set of stock data.

Tech Stack: Python, pandas, numpy , Flask, Postman, pycharm


Titan MIS/GHS Reporting(Jan 2016 – Jun 2016)

Role: Sql Developer/ETL Developer

Project Description: Development and maintenance of MIS system that hosts sales,inventory,after service reports of various business segments.

  • Designed and developed the data warehouse for Titan Management information system.
  • Upgraded existing Crystal reports to SSRS and Power BI reports resulting in 75% decreased runtime of reports.
  • Used SSIS SCD components to capture slowly changing dimensions to generate historically accurate reports.
  • Upgraded existing C# /HTML reports by Converting inefficient Microsoft SQL Queries to Recursive CTEs and efficient Stored procedures.
  • Used Snapshot and Cache mechanism of SSRS to reduce the load on data warehouse.

Supported and maintained existing report portals and data base ecosystem by debugging complex SQL queries and fixing bugs when needed


Tech Stack:SSIS, SSRS, SQL server, C#,ASP.net,Python

Certification

Databricks Certified Data Engineer Associate

Personal Information

Age: 32

Timeline

Data architect

BGSW(Robert Software Global Software )
01.2021 - Current

Data Engineering Specialist

BGSW(Robert Software Global Software )
01.2018 - 12.2021

Data Engineer

BGSW(Robert Software Global Software )
06.2016 - 12.2017

Executive-IT(SQL Developer)

Titan company limited
08.2012 - 03.2015

Executive Trainee-IT

Titan company limited
07.2011 - 07.2012

B.E.(computer science) - Computer Science

Vel MultiTech Engineering College(Anna University
06.2008 - 05.2011

Assistant Manager-IT(ETL Developer)

Titan company limited
4 2015 - 6 2016
Databricks Certified Data Engineer Associate
70-461: Querying Microsoft Sql server
DP100-Microsoft Certified: Azure Data Scientist Associate
Rajasankar RadhakrishnanData Architect/Data Engineer