I joined Capgemini in Apr’21 and have been in the training in cloud technologies for around 3 months and got in the project as a ADF Developer. I'm skilled in working on Azure Data Factory, I used to build pipelines, datasets and dataflows and worked on the optimizations to minimize the run time. With my ongoing experience, I have been validating and the data from the tables and identifying the mismatch in the data if we are having data issues. Resources available on the project learned about code promotion to different environments, creating linked services and modifying the Integration runtimes. Results-oriented Analyst skillful in managing and breaking down large volumes of information. Proactive at heading off issues in operations, workflow and production by uncovering trends affecting business success. Rational Azure Data engineer/Etl Developer with 3.1 years of experience enhancing operations for organizations through information systems solutions. Proven track record of transforming specific business goals related to growth and efficiency improvements into new system designs. Knowledgeable innovator known for recommending new technologies to enhance existing systems and introduce new systems.
In this project, we migrated the data from On-Prem Netezza server to Cloud Azure SQL DB/Azure Synapse/Azure SQL Managed Instance.
My stream of work is on ETL and I'm working as an ADF Developer, coming on to the roles and responsibilities.,
· Need to have a thorough check on the On-prem Informatica structure and had to prepare an analysis sheet for DataMart.
· Must build the model (Dataflows/Pipelines) on Azure data factory based on the On-Prem model (Workflows/Worklets/Tasks).
· Developing the ELT/ETL pipelines.
· Performed Initial/Historical load using csv as source files and informatica DTT task.
· Performing the data validations with the SIT test cases for the build models with a capagemini integrated tool called DVE.
· Performed data validations using DVE tool for count check and data check.
· Scheduling triggers for End-2-End pipelines.
· Monitoring the triggers for success or failure.
· Code movement to Prod, Dev, and QA.using Azure Devops
· Performing tunings to the pipelines, if required.
· Been as a primary spoc for the team.
· Implemented pipeline for scaling SQL DB using ADF whenever pipeline was triggered to improve performance
Project Name :Blue Shield of California(BSC).
Domain : Health Care
Role : ETL Developer(Informatica)
Description:
BSC (Blue Shield of California) is a biggest healthcare system in US.Our project is integrating the data of the patients with their medical details.We extract the data and load into new tables which will be used to send the reminders for future medical tests and analysis
Responsibilities
· Gathered Business requirements and prepared Source to Target Mapping
specifications and Transformation rules.
· Creating Mappings based on the Data Mapping Document.
· Extensively used the transformations such as Source qualifier, Aggregator, Expression, Lookup, Filter, Update strategy, normalizer, and Sequence Generator.
· Creating Sessions, Worklets and Workflows using Workflow Manager.
· Prepared Unit Test Case Document for the mapping developed.
· Involved in code reviewing of the other mappings developed.
Environment: InformaticaPowerCenter 9.6.1,Oracle,Unix,DBeaver
Azure Data Explorer
Azure Data Explorer