Summary
Overview
Work History
Education
Skills
Certification
Projects
Timeline
Generic
AMRITA SINGH

AMRITA SINGH

Data Engineer
Mumbai

Summary

  • 9+ years of industry experience with 7 years in Big data application components like Hadoop, PySpark, Hive, Sqoop, Yarn, HDFS, Map reduce, Python and 2 years in mainframe development
  • Expertise in developing and implementing PySpark scripts using RDD and Dataframe APIs for complex business requirements.
  • Experience creating data pipelines on GAIA private cloud.
  • Worked on Creating AWS glue tables on S3 buckets using crawlers, querying the data using AWS Athena.
  • Created Hive tables and implemented Partitioning, Dynamic Partition, Buckets in HIVE for performance optimization
  • Experience on working with different file formats like parquet, ORC, text for hive and spark
  • Developed Sqoop scripts for importing data to Hive and HDFS from RDBMS, and processed data using Hive-QL
  • Developed data pipelines to read data from data lake on Hadoop, perform complex business transformations and store cleansed and enriched data into Hive tables in Data warehouse.
  • Experienced in creating UNIX shell scripts and SQL.
  • Extensively worked in agile methodology for the project implementation and delivery
  • Good working knowledge in all phases of SDLC in development and enhancement projects which includes requirement gathering, design, development, unit testing, release management, implementation, post implementation validation.
  • Experience in leading, training and mentoring teams, tracking and meeting the deadlines.
  • Effective communication skills and good analytical abilities.
  • Collaborated with end-to-end project managers, technical architects , onshore counterparts, system testing teams, business analysts.
  • Worked on BFSI and Health care domain projects.

Overview

9
9
years of professional experience
1
1
Certification

Work History

Data Domain Architect Associate

JP Morgan Chase
12.2021 - Current

Application Development Senior Analyst

ACCENTURE
12.2019 - 12.2021

Senior Systems Engineer

IBM India
08.2017 - 12.2019

Senior Systems Engineer

INFOSYS
02.2014 - 07.2017

Education

Bachelor of Technology - Electronics And Communication Engineering

Uttar Pradesh Technical University
07.2009 - 2013.07

Skills

    Big-Data Technologies : Spark, Hadoop, MapReduce, HDFS, Hive, YARN

undefined

Certification

AWS Certified developer - Associate

Projects

 

Project: Quant Workforce Optimization

Description: The purpose of Quant Workforce Optimization is to provide insights about various metrics of branches and call centers like the employee absenteeism, required headcount, call center performance metrics. The sources of data include parquet files, flat files and Teradata tables , apply various business transformations and store the data into GOS tables. GOS stands for GAIA Oracle Services which is a cloud offering from oracle on GAIA private cloud. Using this data tableau team creates dashboards to provide various insights.

Responsibilities:


  • Understanding and gathering the requirements by coordinating with business stake-holders.
  • Extracting and loading the data from Teradata, parquet files and csv files into GAIA private cloud.
  • Developing and Unit testing PySpark data pipelines.
  • Coordination with architect, business stake-holders and onshore counterparts.
  • End to end ownership of the project.

Project: Certified Business Entity

Description: The purpose of CBE application is to understand the information that is available about a business as well as understand whether the client has either an existing business relationship, had a previous business relationship or simply interacted with the business in the past. There are various data sources like json files, sql server tables, Teradata tables from which we sqoop the data, cleanse it and apply various business transformations and store the data into SQL server tables from where qlikview team picks the data and creates a dashboard to provide various insights.

Responsibilities:


  • Understanding the client’s requirements and translating it into technical requirements.
  • Extracting and loading the data from SQL server, Teradata sources into hive tables.
  • Parsing and loading json data into hive tables.
  • Developing and Unit testing PySpark scripts for cleansing, transforming and consolidating the data from various sources in a common format.
  • Monitoring and ensuring that the process is optimized for minimal cost to business.
  • I was involved from the requirement gathering phase to the implementation phase of various enhancements.
  • Production issues handling.
  • Creating glue tables on S3 data using AWS crawlers.
  • Mentoring new team members.

Project: Wholesale Analytics

Description: The objective of the project is to build a reporting interface for customer and relationship manager to show the customer growth with respect to their investment, sales, working capital. The data for reporting is taken from financial statements of the customers, transformed and loaded into the target table. These target tables are used by BI team to create the interface using QlikView for reporting purpose.

Responsibilities: 


  • Extracting and loading data using sqoop from various data sources on Hadoop platform in hive table
  • Read the data into Spark RDD, dataframes for various transformations and computations as per the business requirements
  • Processed different data sets files like CSV, parquet from HDFS and hive table into Spark code using python for faster processing of data
  • Did various performance optimizations like using cache, partitioning of the tables in hive and Map Side joins and broadcasting.
  • Involved in creating Hive tables to load the transformed data and stored it in HDFS as parquet format.
  • Performed UIT and SIT for code enhancement or modification as per the business logic requirement
  • Loading data to target Hive table that will be used as input for Qlik interface.

Project: Modernized Migration 

Description: The ACES system is a suite of applications which interface to support the multiple lines of business. The system accommodates variations in state mandates, policies and procedures. Anthem claims system for states Connecticut, Maine and New Hampshire is being migrated from ACES claims which is in the mainframe platform to WGS claims system. WGS is also a mainframe platform. I am involved in analysis, requirement gathering, coding and unit testing, SIT support, release management, post implementation support. 

Responsibilities:


  • Requirement gathering and understanding
  • Analyzing and Modification of Existing COBOL programs.
  • Testing of various components including System Testing, Unit Testing and Integration testing of the products
  • Post deployment support and validation
  • Post deployment defects and new requirement handling
  • Coordination with solution architecture, business analyst, business, testing team, SMEs, project managers
  • Train the new team members with the processes involved.

Project: Staging Database for Anti-Money Laundering Application Client:

Description: Anti-money laundering SDB is an application which receives demographic and KYC data of the American express customers through various upstream systems and maintains a database of the same. It also passes data to the downstream system (Cadence) which rates the individual accounts on risk levels for money laundering. 

Responsibilities:


  • Requirement gathering and understanding
  • Coding and unit testing
  • System testing support and defect resolution
  • Business acceptance testing and defect resolution
  • Implementation and Deployment
  • Post deployment support and validation
  • Post deployment defects and new requirement handling
  • Coordination with solution architecture, business analyst, business, testing team, SMEs, project managers
  • Tracking the milestone dates
  • Coordination with onshore

Timeline

Data Domain Architect Associate

JP Morgan Chase
12.2021 - Current

Application Development Senior Analyst

ACCENTURE
12.2019 - 12.2021

Senior Systems Engineer

IBM India
08.2017 - 12.2019

Senior Systems Engineer

INFOSYS
02.2014 - 07.2017

Bachelor of Technology - Electronics And Communication Engineering

Uttar Pradesh Technical University
07.2009 - 2013.07
AMRITA SINGHData Engineer