Dynamic Lead Data Engineer with a proven track record at Hexaware/IQVIA, enhancing big data pipelines and architectures using Spark, Scala, and Hive. Achieved a 3x speed increase in data processing, demonstrating exceptional problem-solving and collaboration skills. Expert in data warehousing and advanced analytics, committed to delivering high-quality, efficient solutions.
Client: IQVIA
Project: Phoenix
Project overview: US Phoenix is a project that migrates IQVIA's US core business processes to an open-source technology-based big data platform using technologies such as Apache Spark, Scala, Hive, and SQL.
Roles and Responsibilities:
Technologies used:
PySpark, Apache Kafka, Scala, Hive, and SQL.
ApplicaMon:
Business Event Monitoring
The business event is IQVIA's data processing capability language. This is the event source that will provide an overview of what is going on in.
IQVIA managed the enterprise with respect to its data assets and processes. A BEM Logger library is created, which can be integrated to.
any other applica9on. The logger module will generate the unique IDs, construct JSON, and post it to the BEM REST endpoint.
Roles & Responsibilities:
Client: CBA
Project : FCT
Application data migration to Azure:
The on-premise data is transferred to Azure for further analysis and reporting purposes. The data was moved to the ADLS Gen 2 location using the Azure Data Factory service, and transformations are done using the Azure Databricks platform. The Databricks notebooks were created and shared with data analysts for ad hoc requests.
Technologies used:
Azure Data Factory, ADLS Gen 2, Azure Databricks, PySpark.
Application: In-Cycle QC Report Module
The reporting module is designed to generate reports based on various requirements from data analysts and the data management team on an ad-hoc basis. It is a Spark application written in Scala. Reports are being generated from Hive using Spark SQL. Some of the reports include Frequency Distribution, Drill Down Report, etc. The tool will read the metadata and generate a report accordingly. The user needs to set up metadata based on their reporting requirements. This is being used by different asset teams to generate reports on top of millions of records.
Technologies used:
Apache Spark, Scala, Hive, SQL
Application: COAF Cloud Migration
The overall purpose of the project is to assist state regulators in the performance of their regulatory oversight function with regard.
To the insurance industry. To rebuild the report, we are extracting the data from the source CDL layer and performing the transformation.
On all layers generating.
The report, as per the mapping document.
Responsible for the data analysis.
Applica9on: Atlas datalake
New Default Management platform that enables seamless workflow, automated control, and real-time tracking. It will create.
Efficiencies throughout consumer default by eliminating multiple systems and related maintenance, eliminating ancillary databases.
and facilitating multi-product contact strategies. It will enable the use of decision engines to drive practices with greater regulatory.
compliance. Provide the required data from various sources to collection 360 (CM3) application, and also ingest CACS data to ATLAS.
Datalake and load into PDOA database for reporting purposes.
Roles & Responsibilities:
Date of Birth: 02/03/90
10, 6, Azure, Apache Spark, KaJa, SQL, Data warehousing, Data modelling, Python, Scala