Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic
Vinay Chavan

Vinay Chavan

Bengaluru

Summary

  • Data & Generative AI Architect with 14+ years of experience in data warehousing, data engineering, cloud analytics, and AI-driven solutions.
  • Expertise across the Azure ecosystem: Databricks (Spark, Delta Lake), Data Factory (ADF), Data Lake, and Synapse Analytics.
  • Strong background in data modeling (conceptual, logical, physical), big data platforms, and both relational & NoSQL databases.
  • Proven track record in building scalable, high-performance data pipelines using Azure Databricks, ADF, and ETL tools such as Pentaho BI Suite and Talend Open Studio.
  • Migration Leadership & Cost Optimization:
    Led and designed migration from Pentaho ETL to Azure Databricks, modernizing legacy ETL into Spark-based pipelines.
    Migrated Databricks (parquet-based implementation) & Azure Analysis Services to Delta tables & Power BI Premium, significantly reducing costs and improving scalability.
  • Hands-on experience in Generative AI integration, including LLMs, vector databases, and Retrieval-Augmented Generation (RAG) for enterprise workflows.
  • Proficient in developing API-driven applications with Python, FastAPI, and Azure Functions.
  • Pre-sales experience contributing to several proposals and solution designs around Databricks, data modernization, and advanced analytics.
  • Skilled in DevOps practices, including Git-based version control, CI/CD pipeline integration, and collaborative agile delivery.
  • Global exposure with 2.5 years onsite in the USA and project engagements across the UK, ANZ, Chile, Mexico, and Spain.

Overview

14
14
years of professional experience
1
1
Certification

Work History

Gen AI and Data Architect (Customer - J&J)

SyrenCloud
11.2024 - Current
  • Architected and implemented data pipelines using Azure Data Factory and Azure data lake storage to enable robust ETL processes across structured and unstructured data sources.
  • Built and deployed serverless compute components using Azure Functions for pdf embeddings and event-driven processing.
  • Designed and developed scalable APIs using FastAPI and Python to support real-time AI inference and data-driven applications.
  • Integrated LLMs (e.g., OpenAI, Azure OpenAI) into business workflows using vector databases, embeddings, and retrieval-augmented generation (RAG) techniques.
  • Spearheaded the design and delivery of Generative AI solutions to accelerate document processing, customer support automation, and internal knowledge systems.
  • Ensured platform scalability, cost optimization, and compliance with enterprise security standards on Azure.
  • Provided technical leadership, code reviews, and architectural guidance to cross-functional engineering teams.

Solution Architect (Customer - Unilever and Mars)

Saama Technologies Inc.
06.2022 - 09.2024
  • Led the migration of a legacy Trade Promotion Analytics product from Pentaho Data Integration and Tableau to a modern stack using Azure Databricks and Power BI Premium, delivering significant cost reduction and improved scalability.
  • Rolled out the platform to 20+ countries, supporting global and regional sales, marketing, and finance teams with standardized KPIs and advanced analytics.
  • Designed and implemented a metadata-driven architecture, with a custom Web App to manage country-specific metadata, enabling multi-country deployments, reusability, and governance.
  • Built governed Delta Lake layers and enforced fine-grained access control with Unity Catalog, ensuring secure, multi-tenant analytics across regions.
  • Developed robust ETL pipelines in ADF and Databricks to ingest data from ERP systems, syndicated data providers, and partner APIs, ensuring reliability and timeliness of insights.
  • Delivered Power BI Premium dashboards with semantic models, enabling regional trade marketing teams to perform scenario planning, promotional simulations, and store-level ROI analysis.
  • Orchestrated CI/CD pipelines using Azure DevOps, automating deployment of data pipelines, notebooks, and infrastructure-as-code (IaC) components.
  • Collaborated with global stakeholders to standardize metrics, align data models, and drive adoption, resulting in faster decision-making and unified reporting.
  • Optimized compute and storage costs through Delta Lake optimizations, partitioning, and right-sizing compute resources, ensuring sustainable global operations.
  • Provided architectural guidance and mentoring to distributed teams, ensuring consistent delivery aligned with the enterprise data strategy.

Technical Architect (Customer - Unilever and Mars)

Saama Technologies India Pvt Ltd
12.2014 - 05.2022
  • Implemented a Trade Promotion Analytics solution using open-source ETL (Pentaho Kettle) and Tableau, including a custom workbench for admin users to manage configurations and governance.
  • Built governed Delta Lake layers and applied fine-grained access control with Unity Catalog, enabling secure, multi-tenant analytics across 20+ countries.
  • Developed and optimized ETL pipelines in ADF and Databricks, ingesting data from ERP systems, syndicated sources, and partner APIs to ensure timely and reliable insights delivery.
  • Developed a scenario planner Web App supporting predictive analytics and ROI modeling, enabling regional teams to simulate promotional outcomes and optimize trade spend.
  • Designed and built Universal Data Lake and Business Data Lake for the client (Unilever), leveraging Azure Databricks to consolidate diverse data sources and provide a governed foundation for advanced analytics.
  • Led the migration of the legacy platform (Pentaho + Tableau) to a cloud-native stack (Azure Databricks + Power BI Premium), implementing a metadata-driven architecture with a Web App to manage multi-country metadata.
  • Orchestrated CI/CD pipelines in Azure DevOps for automated deployment of data pipelines, notebooks, and infrastructure-as-code (IaC).
  • Collaborated with global stakeholders (Sales, Marketing, Finance) to standardize KPIs, align data models, and drive adoption of the platform globally.
  • Optimized compute costs and improved performance through Delta Lake optimizations and right-sizing compute, delivering measurable TCO reduction.
  • Mentored global delivery teams (onshore + offshore) and ensured architectural alignment with enterprise data strategy.

Senior ETL Developer

Ezest Solutions Pvt Ltd
04.2014 - 10.2014
  • Designed and developed robust ETL pipelines using Pentaho Data Integration (Kettle) to support financial reporting, customer analytics, and regulatory compliance.
  • Extracted and integrated data from MySQL, MongoDB, and third-party financial systems, transforming raw transactional data into structured formats for analysis and reporting.
  • Built reusable transformation templates and parameterized jobs to standardize data loads across different business units.
  • Developed and optimized complex SQL queries, views, and stored procedures in MySQL for data transformation, cleansing, and validation.
  • Modeled and processed semi-structured data from MongoDB, converting nested JSON structures into relational formats for audit and reconciliation.
  • Supported daily batch processing of critical financial data, maintaining SLAs and ensuring zero data loss in high-volume environments.
  • Collaborated with data analysts, finance teams, and auditors to understand business rules and ensure compliance with financial data governance standards.
  • Maintained comprehensive documentation for data flows, business logic, and transformation mappings, enabling transparency and audit readiness.

Senior ETL Developer

Persistent Systems Ltd.
06.2011 - 04.2014
  • Developed and maintained scalable ETL pipelines using both Pentaho Data Integration and Talend to support on-shelf availability (OSA) reporting, inventory tracking, and store-level performance analytics.
  • Integrated data from diverse retail systems (POS, inventory, replenishment, planograms) into SQL Server data warehouses to provide a unified view of product availability across regions and formats.
  • Designed and implemented complex data transformations to align supply chain, product hierarchy, and store layout data for accurate shelf availability metrics.
  • Built job orchestration workflows with detailed logging and error-handling mechanisms to ensure data reliability and consistent SLA adherence.
  • Optimized SQL queries, stored procedures, and indexing strategies within SQL Server to improve ETL performance and support near real-time reporting.
  • Collaborated with category managers, supply chain analysts, and IT stakeholders to refine data requirements and deliver actionable insights for operational decision-making.
  • Documented end-to-end data pipelines, source-to-target mappings, and business logic to support data governance, transparency, and auditability.


Education

B.E. IT. -

Shivaji University
Kolhapur, India
05-2011

H.S.C. -

New College
Kolhapur, India
05-2007

S.S.C. -

Maharashtra
Kolhapur, India
01.2005

Skills

  • Python
  • PySpark
  • Azure Data bricks
  • Azure data factory
  • Gen AI
  • Vector Databases
  • Fast API
  • Data Modeling
  • ADLS
  • Azure Synapse
  • Azure Functions
  • Azure Data Lake
  • Unity Catalog
  • Pentaho Data Integration (Kettle)
  • PowerBI
  • No SQL databases (MongoDB)
  • Databases and SQL

Certification

Databricks Certified Data Engineer Associate

Databricks Certified Gen AI Associate

Timeline

Gen AI and Data Architect (Customer - J&J)

SyrenCloud
11.2024 - Current

Solution Architect (Customer - Unilever and Mars)

Saama Technologies Inc.
06.2022 - 09.2024

Technical Architect (Customer - Unilever and Mars)

Saama Technologies India Pvt Ltd
12.2014 - 05.2022

Senior ETL Developer

Ezest Solutions Pvt Ltd
04.2014 - 10.2014

Senior ETL Developer

Persistent Systems Ltd.
06.2011 - 04.2014

B.E. IT. -

Shivaji University

H.S.C. -

New College

S.S.C. -

Maharashtra
Vinay Chavan