Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic
Sumanth Gaddam
Open To Work

Sumanth Gaddam

Data Enginner
Bangalore,KA

Summary

Senior Data Engineer with 4+ years of experience delivering production-grade, enterprise data platforms using Microsoft Fabric, PySpark, SQL, and Delta Lake. Proven track record of building scalable, idempotent data pipelines, implementing SCD Type 2 dimensional models, and enabling analytics-ready datasets for executive reporting. Demonstrated impact in improving data reliability (25–30%), supporting point-in-time analysis, and accelerating business decision-making through trusted KPI layers. Strong ownership of the end-to-end data lifecycle, with cloud-agnostic design patterns applicable across Azure, AWS, and GCP.

Overview

4
4
years of professional experience
2
2
Certification

Work History

Data Engineer

Hamsa IT Solutions
Bangalore
09.2021 - Current
  • Designed and operated batch data pipelines to convert transactional data into analytics-ready datasets.
  • Developed PySpark workflows with schema enforcement and business-rule validations for data integrity.
  • Implemented incremental processing and historical backfills, enabling safe reruns without duplication.
  • Constructed fact and dimension tables for accurate trend and point-in-time analysis.
  • Delivered Gold-layer datasets that powered executive and operational Power BI dashboards.
  • Reduced pipeline reprocessing failures by approximately 25-30% through idempotent design principles.
  • Processed transactional datasets up to 20 GB per run with production-grade reliability.
  • Applied distributed data engineering best practices across various cloud platforms.

Key Project

Enterprise Banking Data Engineering Platform
Transaction Analytics & Reporting
05.2024 - Current

Problem Statement
Lack of a reliable, scalable data platform to support high-volume banking transaction analytics and historical point-in-time reporting, resulting in reconciliation issues and reduced trust in KPIs.

Solution Provided

  • Designed a Medallion (Bronze–Silver–Gold) Lakehouse architecture using Microsoft Fabric.
  • Built PySpark ingestion and transformation pipelines with embedded data quality and schema validation.
  • Implemented SCD Type 2 customer dimension and analytics-optimized star schema.
  • Developed idempotent Gold-layer pipelines using partition-level delete-and-insert strategies.
  • Added audit, reconciliation, and validation controls across pipeline stages.

Business Outcome

  • Improved data accuracy, consistency, and trust in executive and regulatory reporting.
  • Reduced reconciliation issues and reporting discrepancies.
  • Enabled faster, confident decision-making through reliable KPI dashboards.
  • Established a scalable foundation for future analytics and compliance use cases.

Education

B.Tech -

SASTRA Deemed University
08.2021

Skills

  • Language: SQL, PySpark
  • Data engineering: ETL/ELT, batch and incremental processing, SCD Type 2, star schema, dimensional modeling, idempotent pipelines
  • Platforms and tools: Microsoft Fabric (Lakehouse, pipelines, warehouse), Delta Lake, Power BI, CI/CD (Azure DevOps)
  • Architecture and concepts: medallion architecture (bronze–silver–gold), point-in-time analysis, data quality and validation, pipeline orchestration, distributed data processing

Certification

• Microsoft Certified: Fabric Data Engineer Associate (DP-700), 02/12/25 - CredID 5AC9957D676CD49A
• Microsoft Certified: Fabric Analytics Engineer Associate (DP-600), 23/12/25 - CredID 350CB10440C34B76

Timeline

Key Project

Enterprise Banking Data Engineering Platform
05.2024 - Current

Data Engineer

Hamsa IT Solutions
09.2021 - Current

B.Tech -

SASTRA Deemed University
Sumanth GaddamData Enginner