Summary
Overview
Work History
Education
Skills
Key Technical Skills
Projects
Community Engagement
Activities
Timeline

Harikesh Shinde

AI Systems Engineer
Pune

Summary

Dynamic HPC-AI Application Engineer with a robust history of optimizing and porting scientific applications for peak performance on high-performance computing systems. Expertise in aligning HPC applications with specific workloads to ensure efficient resource utilization across diverse computing environments, enhancing the capabilities of HPC facilities throughout India. Proven ability to troubleshoot complex challenges and deliver innovative solutions that drive operational excellence. Deep knowledge of parallel programming, optimization techniques, and HPC architectures, complemented by proficiency in C/C++ and Python to create effective and scalable solutions.

Overview

8
8
years of professional experience

Work History

Sr Software Systems Design Engineer

AMD
06.2025 - Current
  • Led the design and implementation of software systems, enhancing performance and reliability across multiple product lines.
  • Collaborated with cross-functional teams to integrate advanced features, resulting in improved user satisfaction and engagement.
  • Engaged in continuous system monitoring and debugging, ensuring high availability and performance of critical applications.
  • Participated in technology evaluations and prototype development, driving innovation and supporting strategic planning initiatives.

Senior HPC Engineer

8bit.ai
02.2025 - 05.2025

HPC Infrastructure Setup & Management

  • Design, deploy, and manage HPC clusters, ensuring high availability and performance.
  • Configure and optimize job schedulers (SLURM, PBS, etc.) and resource management tools.
  • Implement and maintain high-speed networking (InfiniBand, NVLink, RDMA).

AI/ML Benchmarking & Performance Tuning

  • Benchmark AI/ML applications on HPC infrastructure using frameworks like PyTorch, TensorFlow, and JAX.
  • Optimize GPU-accelerated workloads across various architectures (NVIDIA A100, H100, V100).
  • Analyze performance bottlenecks and propose optimizations.

Solution Architecture & Customer Engagement

  • Work with end customers to assess HPC requirements and design tailored solutions.
  • Evaluate hardware and software stacks for AI, deep learning, and scientific computing workloads.
  • Provide technical guidance on scalability, cost optimization, and workload balancing.

Networking & Storage Optimization

  • Configure and troubleshoot InfiniBand, NVLink, and high-speed interconnects.
  • Integrate high-performance storage solutions (Lustre, GPFS, NFS, Ceph) with AI workloads.
  • Automation & System Monitoring
  • Develop automation scripts (Bash, Python, Ansible) for HPC cluster deployment and management.
  • Implement monitoring and alerting systems (Prometheus, Grafana, Nagios) for proactive system health checks.
  • Presented technical findings to stakeholders, ensuring clear understanding of project status and goals.

Senior Project Engineer

Center for Development of Advance Computing
01.2023 - 01.2025
  • Optimized and ported scientific applications to run efficiently on high-performance computing systems, including clusters, grids, and clouds
  • Architected and evaluated various hardware architectures for performance optimization, including: Intel Microarchitectures, GPGPUs, Graphcore IPUs, Cerebras HPC Accelerators, MiPhi Memory Solutions, IBM Watson Machine, etc
  • Spearheaded entire benchmarking front for system evaluation with various GPU architecture sticking to HPC and AI domain specific applications under MLPerf consortium
  • Collaborated with researchers to understand application requirements and develop optimization strategies
  • Worked with HPC operations team to ensure seamless integration of optimized applications with HPC infrastructure
  • Participated in code reviews and ensured adherence to coding standards and best practices
  • Mentored junior engineers and provided technical guidance and support.

Project Engineer

Center for Development of Advance Computing
09.2019 - 12.2022
  • Experience to have work exposure for projects under NSM (National Supercomputing Mission) with HPC-AI Technologies
  • Major contribution to research domain PoC works and benchmarking of applications on cluster environment
  • Assisted in setting up and configuring of new HPC cluster, including software setup and system integration
  • Used tools to automate application deployment and testing using tools like SPACK, ReFrame, etc
  • Conducted performance benchmarking and tuning, for improvement in computational efficiency for key research projects
  • Provided technical support and training to 1000+ end-users, enhancing their ability to effectively utilize HPC resources for their computational needs.

Former Intern

E-arth Solutions Pvt. Ltd
12.2017 - 01.2018
  • Researched and identified new technologies and approaches, helping to proactively solve unique problems
  • Analyzed, designed, developed, and tested software, including embedded devices for organization's products and systems.
  • Gained valuable experience working within a specific industry, applying learned concepts directly into relevant work situations.

Education

PG-DBDA -

Center for Development of Advance Computing, Pune, India
Percentage: 71.75

B.E - Electronics and Telecommunication - undefined

Sanjivani College of Engineering, Ahmednagar, India
Percentage: 65.42

Skills

Programming Languages: C/C, Python, Shell

Key Technical Skills

  • Parallel Programming, Application - Porting, Optimization, Benchmarking, SLURM Workload Management
  • CNN, RNN, Transformers, Tensorflow, Pytorch, Deepspeed, OpenCV, NLP, LLMs, MLOps, GenAI

Projects

HPC Projects under National Supercomputing Mission:

Project Summary: The National Supercomputing Mission (NSM) is a government initiative in India aimed at boosting the country's supercomputing capabilities. Launched in 2015, the mission involves setting up a network of high-performance computing (HPC) facilities across India. The goal is to empower researchers and scientists with advanced computing power to tackle complex problems in various fields, such as climate modeling, drug discovery, and materials science.

Responsibilities: Parallelization and Optimization of scientific codes on CPUs/GPUs, Parallelize and Optimize given scientific application on CPUs/GPUs (on single node or multi-node cluster), Work with the domain expert on defining the objectives and outcomes of the Parallelization and Optimizations efforts, Mentoring teams on Parallelization and Optimization efforts for ongoing projects, Training on Parallel Computing technologies.

Roles: Porting, Profiling, Optimization, Debugging and getting benchmarks of scientific applications and checking the application scalability. Providing support to researchers to interface with HPC infrastructures. Implementing and serving the management of job scheduling systems. Maintaining application like GROMACS, LAMMPS, NAMD, OpenFOAM, Quantum Espresso, WRF(Weather Research Forecasting), etc. Exposure of having worked on NSM Facilities pan India with 14+ NSM Sites, total computing power up to 22+ PF

Technical Skills: Parallel Programming, HPC, SLURM Workload Management, Optimization, Containerization, Benchmarking Application 


Artificial Intelligence Projects under National Supercomputing Mission:
Project Summary: The National Supercomputing Mission (NSM) in India aims to enhance the nation's capabilities in the field of HPC and make supercomputing resources accessible to research and industry. Under this mission, the Indian government has been pushing forward various artificial intelligence (AI) and machine learning (ML) projects to boost innovation and provide solutions in critical sectors.
Responsibilities: Developed a real-time image processing based application pipeline using TensorFlow and a GPU-accelerated CNN and Transformers for tasks like computer vision, natural language processing (NLP). Leading the benchmarking activities for evaluating AI application across GPU platforms ranging from different OEMs. Created in house benchmark application kernels for system acceptance test and evaluation. Working with cross-functional teams to deliver AI solutions.
Roles: Develop Machine Learning and Deep Learning models to solve complex problems, Implement advanced AI techniques (e.g., CNNs, RNNs and Transformers) for tasks like computer vision and Natural Language Processing (NLP). Gathering the data for image segmentation and classification task, data annotation, and preprocessing for creating the model. Deployment of ML applications on platforms like HuggingFace, Streamlit, Gradio, Heroku, etc. Actively support NSM sites and internal groups of AI/ML workloads. Benchmarking AI Application on open consortium platform like MLCommons.
Technical Skills: Python Programming, ML, DL, Relevant libraries awareness (Tensorflow, Pytorch, etc), OpenCV, NLP, LLMs, MLOps


PARAM Shavak Product Development:

Project Summary: PARAM Shavak - solution, aims to provide computational resource (Capacity building) with advanced technologies to perform high-end computations for scientific, engineering and academic programs to address and catalyze the research using modelling, simulation and data analysis. This initiative is expected to create HPC aware skilled workforce (Capability building) and for promoting research by integrating leading-edge emerging technologies at grass root level.

Responsibilities: System setup and configuration according to client requirement, Updating the product documentation, Ensuring the integrity with latest technology support. Technical support for client along with training on different variants of PARAM Shavak series like HPC, DL-GPU and VR.

Roles: System Solution Design, Development & Integration, Technical evaluation, System development, Product Documentation, Training activities for the marketing teams, Deployments at the client site.

Technical Skills: Shell & Python Scripting, Product Management


Community Engagement

  • Active member in MLPerf Community which is an engineering consortium, built on a philosophy of open collaboration to improve AI systems and to assess transparent benchmarking process
  • Member of High Performance Software Foundation (HPSF) which is an open-source, vendor neutral hub for high performance software projects, working groups, events, and training
  • Working on Open Research Group for Exascale Computing Project (ECP)
  • Attended conferences and meetups to connect with other HPC professionals and share knowledge

Activities

  • Product Development of Param Shavak and integration of new tools and services
  • Active participation in organizing the Hackathon in collaboration with NVIDIA
  • Completed Certification in Deep Learning Specialization online from DIAT-Pune, and Future Skill Prime.

Timeline

Sr Software Systems Design Engineer - AMD
06.2025 - Current
Senior HPC Engineer - 8bit.ai
02.2025 - 05.2025
Senior Project Engineer - Center for Development of Advance Computing
01.2023 - 01.2025
Project Engineer - Center for Development of Advance Computing
09.2019 - 12.2022
Former Intern - E-arth Solutions Pvt. Ltd
12.2017 - 01.2018
Sanjivani College of Engineering - B.E - Electronics and Telecommunication,
Center for Development of Advance Computing - PG-DBDA,
Harikesh ShindeAI Systems Engineer