Summary
Overview
Work History
Education
Skills
Personal Information
Others
Timeline
Generic

Asif Shaikh

Mumbai

Summary

I am a passionate Data Engineer with over 13 years of experience with deep knowledge of Google Cloud Platform and cloud computing services and Big Data Platform(Hadoop).Extensive experience in design, build, and deploy ETL and ELT data pipelines in the cloud, to ingest data from various sources (SAP,Non-SAP).Good working knowledge on query optimization, indexing, and performance tuning.Excellent communication skills to effectively collaborate with cross-functional teams, including data scientists, analysts, and business stakeholders. Ability to convey technical concepts to non-technical stakeholders in a clear and concise manner. Ability to actively participate in or lead discussions with clients to identify and assess concrete and ambitious avenues for improvement.

Overview

13
13
years of professional experience

Work History

GCP Data Engineer

Mondelēz International
Mumbai
09.2022 - Current
  • Working on building a data ingestion strategy to load data into BigQuery.
  • Responsible for creating ELT pipelines to load the data from different systems onto BigQuery, and building views on sales and prediction data for reporting and analysis requirements on Tableau.
  • Engaging with product owners and stakeholders to understand the requirements and smoothly deliver the project as per the scheduled timelines.
  • As an Aecorsoft SME, extract SAP data and load into GCP using different types of task/config.

Lead GCP Data Engineer

Accenture
Mumbai
01.2021 - 09.2022
  • Worked as Lead data Engineer for the SKY better data project, the project is about migrating all the sky viewing data from Netezza to GCP.
  • Data acquisition to GCP big query table from on-prem.
  • Merge and transform the source based on LLD by writing ETL queries in BigQuery.
  • Extract reports in GCS and send it to downstream users.

GCP Data Engineer

Accenture
Mumbai
12.2019 - 12.2020
  • Worked as a Data Engineer for Now TV / BIT Migration project. The project was about migration the subscription data from Hadoop to GCP.
  • Migrating historical data to bigquery following the ELT approach.
  • With the help of Google orchestration tool Cloud Composer, scheduled the Data Pipelines as per the Business need in the Apache Airflow environment which consisted of various DAGs as per the design and used the Cloud PUBSUB to publish the messages to the Subscribers of the data.
  • Created the CI-CD pipeline process using GitHub and cloud trigger.

Big Data Developer

Lloyds Bank
London
02.2018 - 12.2019
  • As a part of this project all MBNA customers, their demographic data, their email address, Telephone numbers, Product information, Marketing consents and statement preference will be transferred to LBG system through a very complex process of finding matching customers between LBG and MBNA and merging their product information/details.
  • Worked as Hadoop developer in project to deliver the solution which brings customer data from Mainframe systems to Hadoop ecosystem for more than 3.9 million of customers and their demographic data, marketing consents and preferences.
  • Delivered some of the complex logic of customer match and merge process, scrub logic, Masking, Bulk data processing, special character handling in hive.

Big Data Developer

Best Buy
Mumbai
07.2012 - 01.2018
  • This project is a Big Data project and motive is providing real time Online Recommendations to the customers. Development is done using Agile methodology.
  • These recommendations were deployed on BestBuy dotcom, mobile, tablet and were sent through emails.
  • Process large volumes of customer and sales data to show recommendations to end user based on sales patterns and user specific behavior.
  • Involved in creation of Batch jobs using technologies such as Oozie, Apache Pig, Apache Hadoop and Apache Hive to process the data and generate recommendations.
  • Involved in creation of RESTful APIs using Spring MVC and Spring Web to interact with the Dotcom website to display the recommendations to the end user.
  • Data Modelling for Apache Cassandra.
  • Managing the source code version control using Git Stash and Source tree tool.
  • Involved in automated testing of RESTful web services using JBehave testing tool.

Education

Bachelor of Technology - Electronics

MPSTME, NMIMS University
MUMBAI
05.2012

Skills

  • Google Cloud Platform
  • BigQuery
  • Composer
  • GCS
  • Dataflow
  • Cloud trigger
  • Cloud repo
  • Hive
  • Pig
  • HDFS
  • SQL
  • Python
  • Unix shell scripting
  • SAP Data Integrator - Aecorsoft
  • Oozie
  • Airflow
  • Automic
  • JBehave (Behavior-Driven Development)
  • GitHub
  • Jenkins
  • Jira
  • BigQuery optimization
  • Data ingestion
  • ETL development
  • Data migration
  • NoSQL database: Cassandra

Personal Information

Date of Birth: 06/01/90

Others

  • Empaneled interviewer for Accenture lateral hiring for Google Bigquery.
  • Providing GCP trainings to the new joiners as part of Accenture Training program.

Timeline

GCP Data Engineer

Mondelēz International
09.2022 - Current

Lead GCP Data Engineer

Accenture
01.2021 - 09.2022

GCP Data Engineer

Accenture
12.2019 - 12.2020

Big Data Developer

Lloyds Bank
02.2018 - 12.2019

Big Data Developer

Best Buy
07.2012 - 01.2018

Bachelor of Technology - Electronics

MPSTME, NMIMS University
Asif Shaikh