Summary
Overview
Work History
Education
Skills
Certification
Training
Mentoring
Timeline
Generic

Anusha Belagali

Bengaluru

Summary

Results-driven Data Engineer, with an MBA from Indian Institute of Management Bangalore (PGPEM), known for high productivity and efficient task completion. Skilled in big data processing frameworks like Hadoop and Apache Spark, database management using SQL. Proficient in coding in Python and excels in problem-solving, collaboration, and adaptability to leverage technical skills in developing innovative data solutions across diverse environments.

Overview

8
8
years of professional experience
1
1
Certification

Work History

Senior Software Engineer

Walmart
10.2022 - Current
  • Saved ~$90000 annually by migrating the Spark jobs from GCP Dataproc to Serverless
  • Implemented disaster recovery for data pipelines in GCP Dataproc and Serverless, enhancing resilience to approximately 90%.
  • My paper, 'From Cart to Cuisine – Enhancing Retail Experience with an Integrated Meal Planning System and Community Engagement,' was one of the finalists in the Spark Tech Summit, an annual technical summit hosted by Walmart.
  • Reduced ~70% of failures in production by tuning Spark applications
  • Proposed, designed, and implemented the Spark configuration parameter Optimiser—a tool for suggesting the Spark parameters based on the data volume, average run time, and SLA.
  • Saved ~20 man-hours weekly and reduced response times to production failures by working on the API for a failure analysis categorization dashboard that uses Walmart’s internal GenAI tool, powered by Google Gemini, to enable AI-based failure suggestions based on historical failures.
  • Worked on Root Cause Analysis and fix for production failures and incidents to ensure ~99.99% SLA committed to business.

Software Engineer 3

Walmart
07.2021 - 10.2022
  • Saved ~20 man-hours weekly by designing and implementing a Data pipeline reporting system using Fast API, MySQL, HTML, and Airflow scheduler, deployed on GCP
  • Saved ~40 man-hours weekly by designing the application and implementing the API for the Data Pipeline monitoring system using FastAPI, MySQL, Airflow, and Caching
  • Reduced manual efforts by designing the application and implementing an API for file dependency monitoring to track touch files received from upstream of the data pipeline using Python, MySQL, Airflow
  • Reduced the production impact zone radius by working on a Data Lineage API, which tracks the data flow and delays in the data pipelines, and communicates the same to the downstream using Python, and Oracle.
  • Worked on root cause analysis and production fixes for data pipelines for Walmart stores to ensure 99.99% SLA, and reduced production failures.

Production Engineer 1

Yahoo
03.2020 - 07.2021
  • Saved ~20 man-hours weekly by working on an API for an automation tool to track data pipelines scheduled through Oozie using Python, and Flask.
  • Significantly reduced manual efforts by working on Data lineage API based on touch file dependency for Yahoo Ads Data Systems using Python.
  • Automated the failure identification in data pipelines by setting up a log parsing system using Python.
  • Worked on building automation tools and visualization dashboards to help with operations for the Yahoo Ad platform.

Associate Production Engineer

Yahoo
10.2017 - 03.2020
  • Worked on creating, replicating, and maintaining the Yahoo-specific Cloud MySQL platform which hosted databases for applications across Yahoo.
  • Automated the process for onboarding new schemas to Yahoo Cloud MySQL using MySQL (for process control), CI/CD, and Python.
  • Developed a data pipeline monitoring Slack-integrated bot using Python and MySQL as part of the annual hack day organized at Yahoo. The bot was ranked in the top 10 hacks during the company-wide Hackday 2019.
  • Worked on troubleshooting issues in the data pipelines.

Intern

Yahoo
03.2017 - 06.2017
  • Worked on automating and scheduling the daily reports using bash script
  • Worked on a Data Visualization tool tracing the flow of data from source to target using Graphviz, python and JavaScript

Education

MBA - PGPEM

Indian Institute of Mangement
Bangalore
01-2024

Bachelor of Engineering - Computer and Information Sciences

M S Ramaiah Institute of Technology, Bangalore
01-2017

12th -

10th -

Skills

  • Spark
  • Linux
  • Python
  • Shell scripting (Bash)
  • Javascript
  • SQL
  • MySQL
  • Oracle
  • Airflow
  • Hive
  • GCP DataProc
  • GCP BigQuery

Certification

  • Red Hat Certification for System Administrators
  • System Operations on AWS
  • Javascript certification from Udemy

Training

Hopper’s India 2020, Represented Verizon in Grace Hopper’s conference

Mentoring

  • Yahoo, Mentored the new college graduates who joined as interns and collaborated with them to design and develop an HBase capacity addition determining tool.
  • Walmart, Server load monitoring for on-prem nodes

Timeline

Senior Software Engineer

Walmart
10.2022 - Current

Software Engineer 3

Walmart
07.2021 - 10.2022

Production Engineer 1

Yahoo
03.2020 - 07.2021

Associate Production Engineer

Yahoo
10.2017 - 03.2020

Intern

Yahoo
03.2017 - 06.2017

MBA - PGPEM

Indian Institute of Mangement

Bachelor of Engineering - Computer and Information Sciences

M S Ramaiah Institute of Technology, Bangalore

12th -

10th -

Anusha Belagali