

Dynamic and results-driven Data Scientist with over 11 years of IT experience, specializing in Generative AI, Machine Learning, Natural Language Processing (NLP), and Data Quality Engineering. Strong hands-on expertise in Python, machine learning model building, Generative AI development, data governance, and seamless production deployments. Proven track record of leading end-to-end data science solutions that drive innovation and efficiency across the banking, telecommunications, and manufacturing sectors. Skilled in transforming complex data into actionable insights to support strategic decision-making and enhance business performance.
Project: NLP-Based Multi-Agent Chatbot for Project Intelligence.
Project Description: The NLP-Based Project Intelligence Chatbot is an advanced, agent-driven conversational system designed to answer complex natural language queries related to enterprise program and project management data. The chatbot integrates with PTS, JIRA, and CSI systems, enabling unified access to large volumes of project, task, sprint, epic, and application-level information. Using a combination of NLP, generative AI, and multi-agent orchestration, the chatbot interprets user questions, routes them to specialized retrieval agents, and fetches accurate and real-time insights. It can provide high-level summaries, detailed status reports, sprint progress, epic breakdowns, dependency analysis, resource allocation insights, and issue tracking information. The system leverages intelligent summarization models and context-aware response generation to transform raw operational data into meaningful insights. With capabilities such as semantic search, automated aggregation, cross-system mapping, and contextual summarization, the chatbot significantly reduces manual reporting efforts, and helps teams quickly understand project health, risks, and dependencies across multiple tools.
Responsibilities:
Project: Data Lake Attribute Recommender.
Project Description: The Data Lake Attribute Recommender is an intelligent system designed to automatically suggest relevant attributes for analytics, reporting, and machine learning use cases by analyzing the structure, metadata, and semantic patterns of datasets stored in the data lake. Using advanced techniques such as metadata profiling, semantic similarity matching, statistical correlation, and embeddings-based similarity (for text-heavy attributes), the tool identifies relationships between datasets and recommends the most meaningful attributes to users. This reduces manual effort in attribute discovery, improves data usability, and accelerates the development of analytical pipelines.
Responsibilities:
Project: Meta Data Analyzer.
Project Description: The Metadata Analyzer Tool is an intelligent, automated system designed to evaluate, validate, and extract insights from metadata across large, complex datasets. It streamlines data governance by profiling datasets, detecting structural inconsistencies, identifying data quality gaps, and ensuring compliance with enterprise data standards. The tool scans multiple data sources—databases, files, APIs, and data lakes—and generates detailed metadata summaries, such as data types, patterns, null distributions, relationships, schema anomalies, and lineage paths. It enhances visibility into data assets, allowing data engineers, analysts, and governance teams to make faster, more informed decisions about data readiness and quality. Featuring rule-based validation, automated profiling, and configurable quality checks, the Metadata Analyzer Tool reduces manual effort, improves accuracy, and strengthens the overall Data Quality Framework. Its insights help organizations maintain clean, trustworthy, and analytics-ready data.
Responsibilities:
Project: LC Advising Tool - HSN Code Classifier.
Project Description: The HSN Code Classifier is an NLP-driven machine learning system designed to automatically identify the correct Harmonized System of Nomenclature (HSN) code for goods described in TAG 45 of the SWIFT MT 700 Letter of Credit message. TAG 45 often contains unstructured item descriptions that vary significantly in format, terminology, and detail. To solve this, the tool uses a Long Short-Term Memory (LSTM)–based deep learning architecture, trained on domain-specific trade text and historical classification data. The LSTM model processes the sequential patterns and linguistic context within the TAG 45 description, capturing key product attributes, product type indicators, and distinguishing terms to classify the item into one of the 28 active HSN code classes used within the system. By analyzing free-text descriptions, normalizing inconsistent terminology, and learning semantic patterns, the classifier delivers highly accurate HSN predictions, even for complex or ambiguous goods descriptions. The solution reduces manual classification effort in trade finance operations, improves compliance and audit readiness, standardizes item coding across transactions, and accelerates LC processing workflows. The classifier integrates seamlessly with existing trade finance systems, enabling automated HSN validation, duty estimation, risk scoring, and downstream reporting, with minimal human intervention.
Responsibilities:
Project: Capacity Planning Tool.
Project Description: The Capacity Planning Tool is an analytics-driven workforce optimization system designed to accurately estimate employee capacity requirements for end-to-end banking operations. The tool analyzes historical transaction volumes, operational workloads, processing times, seasonal demand patterns, and SLA commitments to determine the optimal number of employees needed across departments, functions, and process queues. By leveraging statistical forecasting, workload modeling, and task-level productivity metrics, the system predicts future staffing needs and highlights potential capacity gaps. It provides granular insights at daily, weekly, and monthly levels, supporting proactive resource planning and efficient allocation of teams. The tool enables operations managers to simulate different business scenarios, evaluate the impact of volume spikes, process changes, or new product introductions, and make informed decisions on hiring, cross-training, and load balancing. It improves service delivery, reduces bottlenecks, minimizes overtime, and ensures compliance with regulatory turnaround times, ultimately enhancing operational efficiency across banking functions.
Responsibilities:
Project: Telecom Customer Churn Analysis.
Project Description: Customer Churn Analysis performs the task of determining whether a customer will stop using services or continue to use the services provided by BT. It provides a better analysis of parameters that would improve customer retention and help clients plan based on such parameters. It also helps customers in designing their offers for new customers. It also helped customers manage location-based inventory.
Responsibilities:
Project: Air Quality Index
Project Description: As a part of this project, the Air Quality Index was to be predicted, where a huge amount of data was to be fetched from multiple sources by the process of web scraping. I was responsible for pre-processing data, analyzing, and providing predictions by applying several machine learning algorithms, as well as testing and validating all machine learning algorithms. Conducted exploratory data analysis on a quick turnaround project involving data manipulation and analysis, and generated valuable insights.
Responsibilities:
Project: Agile Integration Broker
Project Description: AIB (Agile Integration Broker) is BT's strategic order orchestration system. It is a workflow system that is a pass-through system for orders, where the order journey happens through the AIB tool. It is product model-driven for rapid concept-to-market and is developed following a fully test-driven, agile approach using Java and open-source.
Responsibilities:
Natural language processing
Gratitude Award for excellent performance.
Most Valuable Player: For Outstanding Performance and Valuable Contribution in the Team.
Best Team Award: For smooth deployment and handover of deliverable to client.
Appreciated for learning project and new technology (Python and Data Science) in short period of time.
Got appreciation for quickly adoption of Continuous Improvement program for client.
Got 5 times Insta Awards by Peers for valuable contribution Towards Team.
Certified Scrum Master (CSM) by Scrum Alliance
Advanced Learning Algorithms by Stanford and Deeplearning.AI
Supervised Machine Learning Regression and Classification by Stanford and Deeplearning.AI
Machine Learning Specialization by Stanford and Deeplearning.AI from Coursera
Unsupervised Learning, Recommenders, Reinforcement Learning by Stanford and Deeplearning.AI
Certified Data Scientist by VSkills
Certified Scrum Master (CSM) by Scrum Alliance