Experienced technology professional with a proven track record of innovation and adaptability. Skilled in optimizing systems and integrating technical solutions to achieve business objectives. Demonstrated expertise in project management, leading initiatives from start to finish to drive organizational growth and success.
Project Name: FGemini – Ford Domain-Aware Conversational AI (POC)
Description:
Designed and implemented a custom conversational AI assistant ("FGemini") using Google Cloud’s Vertex AI, combining Ford internal domain knowledge (e.g., manufacturing, sales, service) with public data via Gemini Pro and Retrieval-Augmented Generation (RAG). The application mimics a ChatGPT-like interface and intelligently answers cross-domain business queries by integrating structured Ford data with unstructured web sources, demonstrating real-world AI integration at enterprise scale.
Responsibilities and Highlights:
Project Name: Data Platform & Engineering - DP&E
Ford is building a suite of data products tailored to the specific requirements of its portfolio teams. These data products aim to integrate critical Key Performance Indicators (KPIs) into business dashboards, enabling stakeholders to make informed decisions by identifying performance trends and areas that require improvement. The initiative spans across multiple business domains, considering Ford’s 120+ years of operations and extensive data landscape, including legacy systems and historical records stored in various formats.
Tasks and roles which I was involved in:
● Engaged with various portfolio teams to understand their business KPIs, data needs, and the decision-making insights they required from the dashboards.
● Performed in-depth exploration within the enterprise data warehouse to identify relevant data sources; collaborated with source system owners to understand metadata, data lineage, and any transformation logic required for KPI calculation.
● Developed scripts to extract and process the required data for each KPI, and shared sample outputs with stakeholders for validation. In many cases, historical data was maintained in spreadsheets, requiring data ingestion and transformation before integration into the centralized data product.
● Coordinated with multiple domain-specific teams across Ford to gather, validate, and standardize data for consistent reporting and analytics.
● Utilized various GCP services and supporting tools throughout the project lifecycle, including:
Project Name: Ford 3rd Party Data Ingestion and Curation
This project was part of Ford Motor Company's Global Data Insight & Analytics (GDIA) initiative, focusing on the ingestion, standardization, and curation of 3rd-party data sources to ensure consistency, usability, and quality across enterprise-wide analytics platforms. The work involved close coordination between various teams including data stewards, source system experts, and landing/data engineering teams. The objective was to streamline ingestion pipelines, enhance data reliability, and ensure the curated data met analytical and reporting needs.
Tasks and roles which I was involved in:
● Led comprehensive data source analysis by collaborating with data stewards to understand metadata, source-to-target mappings, ingestion timelines, and transformation rules.
● Performed end-to-end assessments of Qlik-based and file-based data sources, distinguishing between direct ingestions into GCP BigQuery and Qlik as a pipeline tool; validated source structures using data dictionaries and communicated findings to data scientists and engineering teams to ensure alignment with downstream requirements.
● Conducted data cleansing, transformation, and validation tasks on ingested data using GCP BigQuery, ensuring consistency and integrity across curated datasets.
● Delivered a Proof of Concept (PoC) using the Qlik tool to demonstrate performance characteristics, assess server utilization, and evaluate its integration into the broader pipeline architecture.
● Built and managed file-based ingestion workflows using a standardized framework that leveraged multiple services such as:
Project Name: Definity
Definity is a migration project like moving the entire project setup from Cloudera to GCP. Here they are using multiple services for their workflow i.e. Control-M, Pentaho with 50 + source systems.
Tasks and roles which I involved:
Project Name: Telecom Argentina
This project is a kind of consulting PSO because things are already developed in the client environment, they are facing
performance issues in their CDF pipelines, and from our side, we are giving possible ways to improve the performance of the
pipeline.
Tasks and roles in which I was involved:
Project Name: Tory Burch
Tory Burch this is a migration project like moving the entire project setup from Cloudera to GCP. Here they are using multiple services for their workflow i.e. Control-M, Informatica, and Azure. Apart from this, they are using dashboards for visualization.
Tasks and roles which I involved:
Project Name: CATO
This application aims to help banks increase their profit by analyzing customer segments and shooting relevant campaigns to keep them engaged. This application was developed using the technologies PySpark, with GCP Cloud.
This application workflow is :
Create campaign->Set filters->Set orchestration->Assign creative->Create contact policy->Run campaign->Capture response.
Allows enterprises to identify profitable customers, profile and segment customers, predict customer response to the communication, treat customers based on profile, communicate with them at a 1-1 level, and track and measure responses.
To create, execute and review the campaigns within a few minutes; track the response of the customer to the communication.
Tasks and roles in which I was involved:
Project Name: NGMI(Next Generation Math Integration)
McGraw-Hill is a book publishing company. SPI Global plays an important role in McGraw-Hill’s business by processing the given raw data file (book’s contents) into a complete book format that can be sold in the Online Market.
Tasks and roles in which I was involved:
Project Name: Company Data Management Platform(CDMP)
This project deals with the centralization and processing of data from various facilities, format, normalizing, and standardizing that data in a way that can be used to create and improve the community inventory records. The main purpose of this project was to move the huge volume of structured data from RDBMS to a big data platform to provide fast query results, high availability, data consistency, and security. This application was developed using technologies such as Hadoop, Spark, Sqoop, Oozie & Hive.
Tasks and roles in which I was involved:
Google Cloud Associate Cloud Engineer
Google Cloud Professional Data Engineer