Staff Data Engineer with 16 years of experience in the areas of Data Engineering, Big Data Processing and Data warehousing. Extensive Experience on building real-time/ Batch Data pipelines, Design and Development of Scalable ETL solutions. Worked on Retail, BFSI, Media and Entertainment domains. Skilled in Spark, Scala, Kafka, GCP, Hive, BigQuery, NoSQL, PL/SQL, UNIX, Python, Hive, Informatica, DataStage, AWS, Adobe Marketing cloud, Looker.
International Data Team -
Currently working as a Staff Data Engineer for CDP/CRM Project in International Data team. Responsible for design, development of building scalable big data solutions. Previously worked with Massmart, Customer Engagement and WCP teams within IDT.
ASDA Data Team -
Worked as a Senior Data Engineer for ASDA Data team. Helped them building solutions in Datalake, data pipeline and Data warehouse.
Realtor.com -
A real estate listings website operated by the News Corp subsidiary Move, Inc. I worked as a Technical lead for the project to built a real-time recommendation model for real estate properties. We were able to succesfully deploy a scalable model in production which handled a data volume of 138 million records. Technologies used - spark/python, Dynamo DB, AWS - EMR/S3/ Athena.
Lloyds' Bank -
Worked for real-time Dynamic Risk Engine Project for Lloyds' Bank as part of open banking standards. Built real-time data pipelines to ingest third party data feeds into the credit score AI System. Responsible for design, development and deploying scalable solutions into production. Technologies used - Spark/Scala , Storm, Kafka, Hbase, Cassandra, Jenkins.
IDIOM -
IDIOM stands for Individual Data Intelligence x omni media which was a product owned by DigitasLBI. Project was to build data pipelines to ingest third party customer feeds and validate/cleanse/transform the data, which then used by the Data science team to provide insights about the customer traffic and navigations which then help them to understand the behaviour and target the customers using relevant ads. Technologies used - Spark/Scala, AWS - EMR/EC2/S3/Lambda/Pipeline/SNS
John Hancock Investments -
Project was to build an enterprise datawarehouse for Mars application - which stores sales/transactional data from all geo locations. Worked as a technical lead. Technologies used - Informatica, Oracle, Sql Server, Unix
Built an enterprise datawarehouse on WWE Network subscriptions data which captures the entitlement, e-commerce, network subscriptions. Build EDW/ datamarts from scratch with reporting layer to generate Daily/ Monthly subscription/ entitlement Dashboards/reports. Lead a migration project to move WWE marketing database from MySQL to AWS Redshift and re- engineered all the ETL pipelines. Technologies used - Informatica, AWS Redshift, MySQL, Windows Powershell/ Batch Scripting
ANZ Bank -
Worked on a data sourcing project, where we processed and transformed mainframe datasets which then integrated into a common platform called Batch Integration Hub. Responsible for design and building of ETL pipelines. Technologies used - IBM InfoSphere Datastage, Oracle, Unix, Control-M
Bank of Montreal -
Built an enterprise datawarehouse for Bank of montreal. Responsible for design and building of ETL pipelines. Technologies used - IBM InfoSphere Datastage, DB2, UNIX
Worked as a .Net developer for an internal project for Human resources to build an Employee and HR module. Technologies used - ASP.Net, SQL Server