I am working as Senior Applications Developer with Telus International, Noida. I am an B.Tech. holder with 4 years of experience in various Big Data Technologies i.e. Azure Databricks, Azure Data factory(ADF), Azure Data Lake, Azure Database, Azure Synapse, PySpark , Python and SQL etc. with a total of 9.5+ years of IT experience.
• Facilitated POCs for Data migration to Azure delivering 100% efficiency.
• Built a data pipeline that ingest data to Azure Data Lake/HDFS from multiple data sources
• Achieved proficiency in Azure Data Factory, Databricks, ADLS Gen2, Delta Lake and other Azure services.
• Led the migration of data from multiple sources to ADLS gen2 thereby empowering teams to analyze it.
• Implemented data quality checks, resulting in a 15% improvement in data accuracy.
• Rather than analyzing data in a database, moved this data to a Data Lake and analyzed it.
• Conducted performance tuning on SQL queries, optimizing data retrieval by 20%.
• Migrated On prem Data Distribution to Azure Cloud Distribution and Hive Data Warehouse using PySpark in less than 1 week.
• Wrote a bunch of transformations in one of the existing PySpark projects.
• Used serverless synapse for one of the daily jobs and optimized it to scan less data by using columnar file formats and other techniques.
Projects:-
(1) Customer 360 Analytics
- Executed data-driven analysis using Azure
Databricks and Data Factory to uncover customer
order patterns, providing valuable insights that
enhanced customer engagement strategies, resulting
in increasing repeat orders count by 30%.
- Led a comprehensive analysis of customer
spending, utilizing SQL queries and data aggregation,
resulting in actionable insights for targeted marketing
efforts and a notable 20% increase in campaign
effectiveness.
- Facilitated strategic decision-making by determining
the number of orders placed by each customer, using
data manipulation, leading to a substantial 25%
increase in customer satisfaction.
(2)Market Analysis
- Orchestrated a Hospitality Domain project,
employing PySpark and Azure Cloud, optimizing
lending processes for new banks and achieving 20%
reduction in processing time.
- Applied data cleaning techniques, resolving 95% of
duplicates, null values, and datatype issues, ensuring
data integrity for downstream teams.
- Implemented Slowly Changing Dimensions (SCD)
strategies, handling updates seamlessly and
enhancing data accuracy by 25%.
• Defined schema and created HIVE tables in HDFS using HIVE queries.
• Created partitions and buckets in Hive to handle structure data.
• Performed data processing such as aggregation, joins, filter as per business rule.
• Created new Services/Operations in Physical and Virtual layer and hosting on API gateway and Service Gateway.
• Exposed Rest web services to the consuming channels.
• Understood Customer enhancement/New Requirement changes and configuration/Customizing the Akana Product.
• Configured custom-policy and existing product policies as per API requirements. Policies – Operational & QoS policies.
• Creating new Services/Operations in Physical and Virtual layer and hosting on API gateway and Service Gateway.
• Exposed Rest web services to the consuming channels.
• Understood Customer enhancement/New Requirement changes and configuration/Customizing the Akana Product.
• Configured custom-policy and existing product policies as per API requirements. Policies – Operational & QoS policies.
• Maintained history and status of issues in the Confluence. Contribute to the improvement of the company’s products, documentation, and processes.
• Configured custom-policy and existing product policies as per API requirements. Policies – Operational & QoS policies.
• Developed the Code as per the requirements.
• Involved in developing the input/view pages.
• Implemented Business logic.
• Wrote UI/Business validations for the owned use cases.
• Unit testing and integration testing of owned use case.