Working as a Hadoop developer to load and transform data using Hadoop ecosystems like sqoop, HDFS, Hive, Impala and spark in a banking and finance domain.
Hadoop Ecosystems
undefinedCertified hadoop and big data expert from Edvancer Edu ventures
Creating data mart using hadoop framework:
Location- Pune
From Jan 2019 - Present
Role: Development and Data validation
Description: Creating a data mart that consolidates lending related data attributes required for generating CECL, CCAR reports. Technology used: Used hive query language for ETL logic implementation. Impala used for analytical needs and Oozie for workflow creation. We also generate extract outputs based on hive transformations applied and send the extracts to downstream consumers in desired file format using spark.
Data transformation using hadoop framework:
Location- Pune
From June 2018 - Dec 2018 (6 months)
Role: Development and Data validation
Description: Working on real estate loans where we transform the data using hive transformations resulting into meaningful calculations used in CCAR and Basel reports. Technology used: Used hive query language for ETL logic implementation. Impala used for analytical needs. Spark for creating data extracts for downstream consumers and Oozie for workflow creation.
Academic project: Dysgrip: Stylus based mobile application for dysgraphia
A stylus model designed to improve handwriting for the students having dysgraphia which will improve the handwriting of the child by adjusting the position of the hand. Applied for patent(Application_id): 201721002078A
Learning project : Weblog analysis
To create an analysis report of a dummy weblog using hadoop framework to identify the top 5 users and top 5 products of the website. Hadoop components used : Hive, Hdfs, Sqoop, and my sql to export and import data