Driven Senior Consultant with 12+ years' experience in analysis, design and developed data engineering programs using Hadoop and Spark ecosystem, Datastage and Teradata to implement business requirements in banking and telecommunication field.
Possesses exceptional interpersonal, problem-solving and analytical skills to provide advice and expertise to client organizations improving business performance.
Overview
12
12
years of professional experience
4
4
years of post-secondary education
Work History
Senior Consultant
Tata Consultancy Services
11.2021 - Current
Responsible for building, maintaining, enhancing sales practice fraud and Negative data solutions
Responsible for Data Collection, Data Transformation, Testing from multiple systems, structuring and organizing and creating automated pipeline to ensure continued workflow
Working with business stakeholders and Data Scientists to help understand requirements and provide solutions
Responsible for Customer Data Analysis using Spark, Spark SQL, python and generate monthly or quarterly reports to present to stakeholders and to automate this workflow
Participating in both on-site and off-site discussions and daily scrum calls for understanding client's requirements and project workflow
Responsible for defining metrics and derived variables, analyze across tables from multiple regions, cross LOB's, extract and consolidate data based on defined metrics
Responsible for improving performance and efficiency of pipelines and building critical Production workflows which will serve multiple client teams who depend on this data for fraud risk calculation and remediation
Building End-to-End Data Engineering pipelines and orchestration
Working with Data Modeling, Data Classification and Reporting teams to build Client specific solutions to save cost and increase efficiency.
Consultant
Capgemini
10.2018 - 10.2021
Batch Data Ingestion from different sources like Vision+ Files and RDBMS etc into data lake using IBM Infosphere Datastage
Use HUE as Web interface for analyzing data in Hadoop
Create HLD/LLD/Source to Target Mapping Documents for moving data from source to process to publish zones
Set up Big Integrate environment for Datastage jobs to run on Hadoop Cluster
Create Hive tables using Beeline commands on Hadoop Edge Node
Extensively used Join, Merge, Lookup, Surrogate key, Change Capture, Funnel, Sort, Transformer, Copy, Remove Duplicate, Filter, Pivot and Aggregator stages for various transformations needed on data to conform to business rules
Develop sequence jobs proper job dependencies, job control stages, triggers and notifications
Create HDFS scripts for copying files, data file manipulation, schedule jobs, data validation etc
Document ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing.
Consultant
Accenture Solutions
07.2016 - 10.2018
Designed and developed ad-hoc ETL solutions, which involves ad hoc complex SQL queries
Interacted with business users on regular basis to understand, consolidate and analyze requirements and present design results with performance scope analysis
Created ETL architecture and provided strategy related to data cleansing, data quality and data consolidation Profile and code logic with proper granularity
Involved in development of FRD (Functional Requirement Document) and FDD (Functional Design Document) Using Teradata utilities like fast load Mload and database partitioning to improve performance
Creating Tableau dashboards for insight based decision
Presenting findings and data to client to improve strategies and operations
Extracting, compiling and tracking data, and analyze data to generate reports
Performing daily data queries and preparing reports on daily, weekly, monthly, and quarterly basis
Managing Vendors from Development to Implementation Activities
Responsible for Design and Solutions to Maintain Central Reference Data Management.
Developer
Wipro
11.2012 - 07.2016
Responsible for gathering requirements and designing data mart entities and data Warehouse Worked closely with Project Team and SME to clarify business requirements and translate the requirements into technical specifications
Significantly involved in new design and business logic implementation
Implementing best practices to ensure optimal performance
Reviewed documentation, peer code and conducted code walk through.
Education
Bachelor of Technology - Computer Science
Amity University
Noida
03.2008 - 05.2012
Skills
Hive/HDFS
Pyspark/SparkSQL
Teradata
Datastage
Autosys
HUE/Beeline
Oracle/RDBMS
Python
UNIX
Consulting Skills
Effective Communication
Techno Functional
Team Building
Detail Oriented
Analytical Skill
Problem solving
Adaptability
Accomplishments
Teradata Certified in year 2017.
Collaborated with team of 8 in the development of Genesis project for Citibank at experience of 4.5 years.
Streamlined code optimization process by creating and executing complex scenario, resulting in 15% better performance.
Assistant Delivery Manager at Tata Consultancy Services, Global Shared ServicesAssistant Delivery Manager at Tata Consultancy Services, Global Shared Services