Summary
Overview
Work History
Education
Skills
Carrer Objectives
Declaration
Timeline
Generic

Sirisha Sanku

Bangalore

Summary

  • A result-oriented engineer with 4 years of experience in design and development of Extraction
    Transformation Loading (ETL) methodology for supporting data transformations and processing data in
    a corporate wide ETL Solutions using Spark on Databricks, ADF, Talend and SQL.
  • Experience in design and development of Extraction Transformation Loading (ETL) methodology for
    supporting data transformations and processing in a corporate wide ETL Solutions using on different
    source systems like mySQL, Oracle, Hive, flat files, XML, HDFS and JSON files.
  • Developed reusable PySpark code using Databricks Notebooks for creating complex Data Pipelines
    that helps in extracting, cleansing, transforming critical data into Lakehouse.
  • Hands on experience on reusable components Joblets in Talend.
  • Worked on Big Data Talend Components.
  • Experience in enhancement on existing ETL Logics, Data loads.
  • Good understanding on Data Modelling using Star schema model, Snow-Flake model on fact and
    dimension tables.
  • Hands on experience on admin activities of Databricks workflows, Talend Administration center
    (TAC) and TMC.
  • Actively participated in the Migration process from lower to Higher environment.
  • Skilled in identifying issues and risks in a timely manner, developed/implemented appropriate mitigation and contingency plans.
  • Implemented SCD logic in Talend, PySpark.
  • Worked under LINUX Environment.
  • Followed Best practices in Talend for good performance of the Tool and Data Base.

Overview

6
6
years of professional experience

Work History

Data Engineer

Mindera
06.2023 - Current
  • Implemented Migration from Data Warehouse to Azure environment, i.e, converted existing IBM DB2 Data stage jobs to Databricks PySpark Notebooks.
  • Created required data pipelines in Data Platform environment.
  • Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability.
  • Contributed to internal activities for overall process improvements, efficiencies and innovation.
  • Extended support to perform Testing activities.

Data Engineer

TCS(Payroll - MBIT)
08.2022 - 11.2022
  • Understand the flow and make necessary changes in Databricks.
  • Migration of jobs from Talend to Databricks.
  • Worked on tuning the performance of Databricks pipelines.
  • Resolving errors triggered in code for jobs in Databricks.

Data Engineer

Cognizant
01.2020 - 07.2022
  • Designed and developed Extraction Transformation Loading (ETL) methodology for supporting data transformations and processing in corporate wide ETL Solutions using PySpark with source as Azure Blob storage and target as Delta tables in Databricks.
  • Created Reports and Dashboards which ran on daily basis and weekly.
  • Job migration from DEV to QA.
  • Job scheduling and Monitoring was done through Databricks webApp.

ETL Engineer

Infosys
02.2018 - 01.2020


  • Understanding the business requirement documents and timely collaborating with the business team to lead the development effort.
  • Developed the ETL code base and helping the Team on the development effort.
  • Developed dynamic job which pulls data from all the source tables and loads into intermediate location.
  • Implemented SCD.
  • Developed Talend ETL jobs by using Oracle components, tFileList, tFileInputExcel, tFilePositional, tFileDelimited, tFileOutput components, tMap, tHash, tBuffer and maintaining standards by ensuring single place to initiate connections, maintaining context groups for dev, prod, utilizing Talend integration Suite features like tParallelize to run multiple threads in parallel.
  • Worked on the reusable components Joblets to reduce the manual work.
  • Hands on Experience on TAC (Talend Administrator Center) related activity such as Scheduling and monitoring of Talend jobs.
  • With collaboration with the ETL admin team created JSON files which then helpful creating the replica of TAC jobs with change in parameter values, in order to reduce the creation effort.

Education

B.E/B.Tech - Computer Science And Engineering

Sri Vasavi Engineering College
PedaTadeplli
05.2017

Skills

  • Azure
  • PySpark on Databricks
  • ADF
  • SQL
  • Talend
  • Python

Carrer Objectives

To secure a challenging Data engineer role in a reputable organization where I can leverage my
expertise in designing and deploying scalable and fault-tolerant data pipelines that support real-time
data processing and analytics. My objective is to help organizations transform their data into actionable
insights that drive business growth and profitability, while constantly improving the performance,
reliability, and security of the data infrastructure.

Declaration

I hereby declare that all the information furnished in this application are true and correct to the
best of my knowledge and belief.

Timeline

Data Engineer

Mindera
06.2023 - Current

Data Engineer

TCS(Payroll - MBIT)
08.2022 - 11.2022

Data Engineer

Cognizant
01.2020 - 07.2022

ETL Engineer

Infosys
02.2018 - 01.2020

B.E/B.Tech - Computer Science And Engineering

Sri Vasavi Engineering College
Sirisha Sanku