Summary
Overview
Work History
Education
Skills
Timeline
Generic

Sujan Marlapalli

Visakhapatnam

Summary

Strategic and results-driven Senior Data Engineer/Analyst with hands-on experience in designing, implementing, and optimizing end-to-end data solutions. Adept at leveraging advanced analytics, cloud-based technologies, and ETL processes to transform raw data into meaningful insights. Proven expertise in managing large-scale data infrastructure, implementing data governance frameworks, and driving data-driven decision-making across organizations

Overview

7
7
years of professional experience

Work History

Senior Data Analyst

Dxc Technology
Bangalore
11.2023 - Current

Azure Data Engineer

Lince Soft Solutions Private Limited
06.2017 - 06.2023

Project 1: Migration from Snowflake Platform to Azure Synapse

Client: Microsoft Direct - L3Harris

Role: Azure Data Engineer

Tools and Technologies:

  • Azure Synapse Analytics
  • Delta Lake
  • Azure Synapse Dedicated SQL Pool
  • ADLS Gen 2
  • Synapse Notebooks
  • Synapse Pipeline
  • Py-Spark
  • Spark SQL

Client Description:L3Harris Technologies, Inc. is a leading American technology company specializing in defense solutions, information technology services, and wireless communications. The company provides a range of products, including command and control systems, tactical radios, avionics, and electronic systems, serving government, defense, and commercial sectors.

Responsibilities:

  • Converted Snowflake DDLs of Tables, Views, SPs, and Semi-structured Tables Scripts to Azure Delta Tables for execution in the Azure environment.
  • Parameterized the Schemas and ADLS location in the DDL scripts for Tables in both Bronze and Silver layer Schemas.
  • Created over 130 DELTA tables in ADLS Gen 2 for both RAW and RPT Schemas.
  • Developed Logging and Error Handling Notebooks for Pipeline Orchestration.
  • Integrated over 60 Stored Procedures Notebooks into the pipeline framework.
  • Established more than 70 Views and 60 Tables in Azure Dedicated SQL pool for the Gold layer.
  • Created Snapshot Notebook for the Brown and Silver layers for restoring Tables Data.
  • Orchestrated the deployment of Main and Child pipelines for data migration across Bronze, Silver, and Gold layers, integrating Stored Procedures and Logging Notebooks.
  • Implemented Pipelines for extracting data from the cloud to ADLS Gen 2, extensively working on copy activities and incorporating Error Handling concepts.
  • Utilized Azure Data Factory activities such as Lookups, Stored procedures, if condition, for each, Set Variable, Append Variable, Get Metadata, Filter, and Wait.
  • Configured logic apps for email notifications to end-users and key stakeholders using web services activity.
  • Stored processed and raw data in Azure Data Lake Gen2 containers.
  • Stored processed data in Azure Synapse Analytics.

Project 2: Huntsman BOLT

Client: Huntsman Corporation, USA

Role: Azure Data Engineer

Tools and Technologies:

  • Azure Databricks
  • Azure Blob Storage
  • ADLS Gen 2
  • Azure Synapse Analytics
  • Azure SQL
  • Azure Logic Apps
  • Spark SQL

Responsibilities:

  • Led the offshore team as an Offshore Lead, overseeing end-to-end project implementation.
  • Implemented a comprehensive project involving data extraction from on-premises into ADLS Gen2, pre-processing using Azure Databricks, and storage in Azure Data Warehouse using ADF for further processing.
  • Set up Integration Runtimes (IR) to connect to SAP and non-SAP systems for data extraction into Azure through ADF.
  • Processed data through Databricks using Spark and Python, storing both processed and raw data in Azure Data Warehouse.
  • Presented processed data reports through Tableau.
  • Created over 20 Tables in Azure Data Warehouse to store specified data.
  • Scheduled triggers in ADF for daily pipeline runs, ensuring regular data updates.
  • Tracked, processed, and updated specific tables daily, ensuring timely deliveries of action items.

Project 3: FEA (Fraud Enterprise Aggregation)

Client: Standard Chartered Bank

Role: Database Developer

Tools and Technologies:

  • SQL Server Management Studio (SSMS)
  • SQL DB

Responsibilities:

  • Used SSMS to work with Microsoft SQL Server, involved in Business requirement document walkthrough to understand functionality.
  • Contributed to understanding System Design and Database Designs.
  • Created SQL Databases, Tables, indexes, and Stored Procedures based on PO requirements.
  • Developed User-defined Functions and DB objects.
  • Involved in Bug Fixing and root cause Analysis.
  • Participated in Performance tuning.
  • Planned and developed ETL pipelines using ADF.
  • Extracted data from Azure Data Lake Storage to Azure Data Factory.
  • Prepared transformations using activities and data flows as per requirements.
  • Scheduled and monitored pipelines, loading transformed data from ADF to Azure SQL data.

Education

JNTU Kakinada University
01.2016

Skills

  • Azure Analytics Services:Hands-on experience with Azure Data Lake Store (ADLS), Azure SQL DW, Azure Data Factory (ADF), Azure Databricks, etc
    Excellent knowledge of ADF building components, including Integration Runtime, Linked Services, Data Sets, Pipelines, and Activities
  • ADF Development:Designed and developed an audit/error logging framework in ADF
    Implemented dynamic pipelines for extracting multiple files into multiple targets using a single pipeline
    Automated execution of ADF pipelines using Triggers
  • Azure Databricks:Proficient in Azure Databricks (ADB) Spark-Python and Spark-SQL
    Experience with Spark RDDs and Data Frames
  • Big Data Concepts:Familiarity with big data concepts, including storage and computing frameworks
  • Pipeline Orchestration:Orchestrated data integration pipelines in ADF using various activities such as Get Metadata, Lookup, For Each, Wait, Execute Pipeline, Set Variable, Filter, Until, etc
  • Administration and Management:Knowledge of basic admin activities in ADF, such as providing access to ADLS using service principles, installing Integration Runtimes, creating services like ADLS, logic apps, etc
    Managed data recovery for Azure Data Factory Pipelines
    Monitoring and managing Azure Data Factory
  • Data Ingestion and Integration:Designed and developed data ingestion pipelines from on-premise to different layers in ADLS using Azure Data Factory (ADF V2)
    Utilized Logic App Service to send email notifications for the success and failure of ADF pipelines
    Extensively used Azure Key Vault resource in ADF Linked Services
    Experience with the integration of data from multiple sources
  • ETL and Data Processing:Extensively used ETL methodology for supporting data extraction, transformation, and processing from sources like Oracle, SQL Server, and files into Azure Data Lake Storage
  • Code Deployment and CI/CD:Deployed code using CI/CD processes
  • General Skills:Data Warehousing
    Data Compiling
    Requirements Analysis
    Technical Writing
    Business Analysis
    Data Modeling
    Data Mapping
    Strategic Planning
    SQL and Databases
    Business Intelligence Systems

Timeline

Senior Data Analyst

Dxc Technology
11.2023 - Current

Azure Data Engineer

Lince Soft Solutions Private Limited
06.2017 - 06.2023

JNTU Kakinada University
Sujan Marlapalli