Data Engineer with 5 years of experience in developing efficient ETL pipelines. Proficient in Apache Spark, Scala, and Java for large-scale data processing and analytics. Delivered reliable solutions that enhance data integrity and performance in distributed systems. Committed to leveraging data for informed business decisions and improved operational efficiency.
-Received start of the month award in oracle
-Got Rishab Award in UHG
-Developed and owned new project in UHG
UHONE-NCP-Project
-Developed microservice to transfer stored data from Kafka topic to topic Reduced manual effort of
finance to apply for failure refunds
- Developed new ETL jobs to extract data and store it in the database, create files and error reports
- Reduced manual effort of finance to apply for failure refunds
Digital Ad Analytics
Tasks/Achievements
- Got real time stream impression data in Kafka queues
- Developed Spark Streaming Code to get data from kafka and wrote transformations in Spark
Fusion data Intelligence (FDI)
Tasks/Achievements
- Built and configured Spark/Scala/Java pipelines to extract, transform, and load massive datasets from
various Fusion Cloud modules (ERP, HCM, SCM, etc.)
- Leveraged Oracle’s prebuilt data foundation and semantic models to reduce manual mapping effort
Enabled incremental and full data refresh patterns, ensuring zero-downtime access and timely insights
- Tuned pipelines for optimal performance and fault tolerance Integrated validation routines to compare
source and loaded data, detect discrepancies, and maintain data integrity across distributed systems
- Extended the data model by integrating custom external sources beyond Oracle’s core modules using
Oracle Data Integrator and OCI connectors