Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

Shivam Shukla

Cloud and Big Data Engineer
Hyderabad

Summary

  • 7.5+ years of total experience in ETL Projects with Big Data technologies
  • 2 years’ experience in Snowflake using Snow-park with AWS services S3, glue and lambda.
  • Analyzing Semi structured data and Structured data on HDFS with Apache spark and Hive.
  • Hands on experience in data migration from Hadoop to Snowflake using snow park API.
  • Worked on SQL and Spark SQL queries to perform ETL and reduce defects.
  • Enhancing Hive query scripts to apply transformations on managed or external tables.
  • Hands on experience in Python Programming and debugging with Spark.
  • Training and consulting the resources in Big Data Hadoop Ecosystem and migrations.
  • Good experience in Automated ETL validations.
  • Good Knowledge in SQL and Python applications.

Overview

7
7
years of professional experience
4
4
years of post-secondary education
7
7
Certifications
3
3
Languages

Work History

Senior Software Engineer

Carelon Global India Solutions [Former Legato ]
Hyderabad
08.2019 - Current

PROJECT PROFILE:

Client: Elevance Health Care, US.

LOB : Health Insurance

Role : Big Data Analyst and Cloud migration.

Environment: Hue Editor, Hadoop, Spark, Hive, Python, Scala.

Project Synopsis:

Elevance is the largest health care provider in US. I am a part of Elevance's NCCT project stands for National Consumer Cost Tool. The objective of this project

is to estimate the cost of all procedures and treatments based on historical claims data. The final estimates are published in the public server for US citizens to

estimate the total treatment costs for any hospital.

Responsibilities:

  • Delivered exceptional client support by promptly addressing concerns and implementing requested changes or enhancements to software solutions.
  • Trained and mentored junior developers and engineers, teaching skills in big data and cloud to improve overall team performance.
  • Work on innovations and automations.
  • Stayed current on industry trends, incorporating emerging technologies into new projects when appropriate one such involving open-AI models.
  • End-to-end project delivery starting from table loads to generating extracts and reporting.
  • Analyzing and validating data loaded into HIVE table and Snowflake tables.
  • Establish file to table loads from AWS to Snowflake tables.
  • Author Glue jobs to execute spark scala code via snowpark.
  • Performing data quality checks in close to 50 tables which are loaded using files.
  • Preparing unit test cases to capture all scenarios as per business requirement.
  • Preparing SQL queries to validate cost estimates are correctly populated.
  • Development and modifications to spark-SQL and shell scripts.
  • Analyzing and debugging errors during jobs execution in Snowflake environment.
  • Analyzing detail reports via Tableau.

Associate Consultant

Capgemini India Private Limited
Hyderabad
11.2016 - 08.2019

PROJECT PROFILE:

Client : Philadelphia Insurance Companies

LOB : Insurance

Role : Hadoop Analyst

Environment : MS SQL Server, Hadoop, Spark, Hive, Python

Project Synopsis:

Philadelphia Insurance Companies (PHLY) designs, markets and underwrites commercial Property/Casualty and Professional Liability insurance products, incorporating value-added coverages and services for industries. The objective of this project was to validate, parse, process and store Structured Data and XML files coming from 6 source systems to generate load ready extract files which could be consumed by Datahub and Info center layer for further processing.

Responsibilities:

  • Authoring SQL queries for transforming Structured data from RDBMS as per ETL specification document
  • To perform various validations on transformed data
  • Authoring SQL queries to extract data from XML files and validating them as per requirement
  • Creating Hive tables to load XML data files and applying transformations on them using Spark-SQL
  • Authoring Py-Spark code to extract source history data from RDBMS and ingest to HDFS
  • Validating tables loaded into HDFS with source tables in RDBMS
  • Automating execution flow and generating required output in CSV form using Spark Automation Script
  • Validating CSV files generated in Development and QA environment
  • Analyzing and debugging errors during jobs execution in Hadoop environment.

Education

Bachelor of Science -

GMR Institute of Technology
Andhra Pradesh
04.2012 - 04.2016

Skills

Tools : Spyder, Vi Editor, Hue, ICEDQ

undefined

Certification

  • ISTQB

Timeline

• Snowflake Snowpro Certified

03-2024

• Google Associate Cloud Engineer

12-2023
  • HackerRank Python Certified (Basci)
10-2021
  • HackerRank SQL Certified (Intermediate)
07-2021
  • Apache spark with Python from Udemy
02-2020

Senior Software Engineer

Carelon Global India Solutions [Former Legato ]
08.2019 - Current
  • Big Data and Hadoop from Dataflair
11-2017
  • ISTQB
04-2017

Associate Consultant

Capgemini India Private Limited
11.2016 - 08.2019

Bachelor of Science -

GMR Institute of Technology
04.2012 - 04.2016
Shivam ShuklaCloud and Big Data Engineer