Summary
Overview
Work History
Education
Skills
Certification
Accomplishments
Currentemployer
Timeline
Generic

SANDEEP REDDY SYAMALA

Hyderabad

Summary

  • I have around 2.5+ Years of Experience in implementation of Data warehouse, Database applications mainly Extraction, Transformation and Loading process.
  • Good Hands-on experience on GCP services and creating efficient data pipelines using GCP services.
  • Good hands-on experience in developing Python scripts, integrating Cyberflow for data pipeline performance monitoring, and utilizing SonarQube to ensure code quality and security.
  • Experience in developing stored procedures and optimizing their performance, including SQL query tuning and Strong Knowledge of SQL and relational database models.
  • Experience in ETL processes, dimensional data modelling (Star Schema, Snowflake Schema, FACT& Dimension Tables), OLTP & OLAP.
  • Experience in design & development of database solutions for enterprise applications using databases like MS SQL Server, PostgreSQL.
  • Experience in loading slowly changing dimension tables.
  • Experience in loading data, debugging mapping, performance turning
  • Good communication and interpersonal skills, ability to learn quickly, good analytical reasoning and high compliance with new technologies and tools.

Overview

2
2
years of professional experience
1
1
Certification

Work History

Data Engineer

HSBC
Hyderabad
11.2023 - Current

Project Name: C48 Collection Management

Technologies: Python, Apache Beam, GCP Workflow, GCS Bucket, Airflow, Pub/Sub, PostgreSQL, Jenkins Pipelines, PyTest, Git.

Description: C48 is a program to build a Generation Collection platform. The recommended collection system is an online solution that supports the collection of delinquent and over-limit customers. It maintains up-to-date collection information of accounts in its database and communicates directly with the accounting system, eliminating the need for paper flow. This allows collection departments to more accurately track and forecast potential volumes each day, prioritizing their work more effectively and efficiently. The Collections and Recovery HASE service categorizes and manages accounts that require special handling. HASE Collections primarily manages past-due accounts but can also track, analyze, and recover other account groups, including VIPs, lost or stolen cards, over-limit accounts, and bankruptcy cases. It offers flexible collection strategies driven by control record parameters. The service interfaces with AR and dialer systems. If the system fails for any reason and becomes unavailable, there is no direct impact on customers. However, collectors will be unable to update any data or collect payments from customers during the unavailability period. From the bank's perspective, this could result in a delay in collecting payments.

Responsibilities:

  • Designed and implemented scalable ETL pipelines using Python and Apache Beam on Google Cloud Platform (GCP) to process and transform large datasets.
  • Managed and optimized data storage solutions in PostgreSQL, ensuring data integrity, query efficiency, and performance tuning.
  • Utilized GCP services such as Cloud Dataflow, BigQuery, Pub/Sub for event publishing, and Cloud Storage for efficient data processing, transformation, and storage.
  • Automated data workflows using Jenkins pipelines for continuous integration and deployment, optimizing data pipelines for real-time and batch processing to reduce data latency, and improve throughput.
  • Developed unit test cases for data pipelines, and implemented continuous integration practices to ensure data accuracy and reliability.
  • Integrated Cyberflow for monitoring data pipeline performance, and used SonarQube to ensure code quality and security by identifying and mitigating vulnerabilities in data processing.
  • Debugged and resolved technical issues; also participated in code reviews and followed coding standards.
  • Collaborated in agile teams, and participated in sprint calls.
  • Tested and deployed applications in production environments.
  • Wrote and maintained technical documentation.

Junior Data Engineer

Staples
07.2023 - 11.2023

Project Name: Custexo Migration

Technologies: Azure SQL, IICS, Oracle, stored procedures.

Description: A leader in workspace products and solutions for over 35 years, Staples has thousands of experts dedicated to applying their intuition, expertise, and experience to bring out the best that technology such as AI has to offer. These experts make work easier, smarter, and more efficient for businesses of all sizes. The company operates in North America through eCommerce and direct sales and is headquartered near Boston, Massachusetts.

Responsibilities:-

  • Successfully migrated over 25 stored procedures, tables, triggers, functions, and indexes from Oracle to Hyperscale using the SSMA tool as part of the migration process.
  • Defining the ETL strategies and providing solutions for creating end-to-end data pipelines using Store procedures
  • Responsible for end-to-end documentation, such as ETL processes in stored procedures, testing, and run manuals, etc.
  • Coordinate and integrate change requests to design.
  • Responsible for executing code reviews.
  • Implemented IICS to extract data from on-premises sources, target data, and load it into the Hyperscale compression tables to validate data between Azure SQL and Oracle.
  • Ensuring proper testing, debugging, and verification of software before deployment.
  • Identifying the performance issues and providing better solutions.

Junior ETL Consultant

Charles Schwab Corporation
07.2022 - 06.2023

Project Name: CDW Migration.

Technologies:IICS, PowerCenter, Big Query, Teradata.

Description: The Charles Schwab Corporation is an American multinational financial services company. It offers banking, commercial banking, investing, and related services, including consulting and wealth management advisory services, to both retail and institutional clients.

Responsibilities:-

  • The Informatica team has converted all 8,000+ workflows from PowerCenter to IICS. The main challenge is to test the workflows in IICS. As the syntax is changed from Teradata to BigQuery.
  • Developing a new workflow by converting existing SQL syntax to the IICS mapping. To have clear visibility to everyone.
  • Testing workflows in a new cloud environment and matching the records process in the new flow with the existing on-prem flow. Any disturbance in the data must be analyzed, and the code should be fixed.
  • For matching the records, we used an internal validation tool that will do an apple-to-apple match and provide a clear mismatch report in each column.
  • Involved in analyzing the count mismatch and data mismatch, and fixed the IICS and BigQuery code as per the analysis.

Education

MBA - Data Science

Koneru Lakshmaiah University
Vaddeswaram Guntur District

Skills

  • SQL, Python, GCP, Apache Beam
  • SQL (Stored Procedures, triggers, functions, and indexes)
  • Databases (PostgreSQL, MS SQL Server, Teradata)
  • GCP (Dataflow, BigQuery, Pub/Sub, GCS, Cloud SQL, Airflow Composer, Cloud Functions)
  • Data Modeling (Star Schema, Snowflake Schema, Hybrid Schema, SCD1, SCD2)
  • ETL tools (Informatica Cloud - IICS)
  • Requirements elicitation, design, and solution documentation, Agile, MS Excel, and PowerPoint

Certification

Google Cloud: Certified Professional Cloud Data Engineer, 07/2025

Accomplishments

  • Received Rockstar performance award for Q2 2024.
  • Received Best technical support award for Q1 2023.
  • Received Best team member award for Q4 2022.

Currentemployer

Bilvantis Technologies LLP, Hyderabad, Telangana

Timeline

Data Engineer

HSBC
11.2023 - Current

Junior Data Engineer

Staples
07.2023 - 11.2023

Junior ETL Consultant

Charles Schwab Corporation
07.2022 - 06.2023

MBA - Data Science

Koneru Lakshmaiah University
SANDEEP REDDY SYAMALA