Summary
Overview
Work History
Education
Skills
Certificates
Timeline
Generic

Aravind Kumar Anugula

Banglore

Summary

  • Working as Data Engineer for the last 10 years, with overall 17 years of IT experience which includes Design and Implementation of Micro Services and Large Scale Distributed Systems Development.
  • Excellent Leadership, Technical, Communication, Analytical and Problem Solving skills and ability to get on well with people including cross-cultural backgrounds.
  • Provide technical thought leadership on Big Data strategy, adoption, architecture and design, as well as data engineering and modeling.
  • Expertise in architecting Big data solutions using Data ingestion, Data Storage
  • Strong background of big-data stack like Spark, Hadoop (CDP, EMR), Kafka, MapReduce, Hive, Snowflake, Sqoop, Airflow, Argo, Oozie, MongoDB.
  • Hands on experience in AWS cloud computing (EC2, ELB, ASG, Lambda, S3, RDS, Cloud Watch, EMR, Data Pipeline, VPC, Route 53, Kinesis, Athena).
  • Proficient in Scala, Java, Python.
  • Experience with Kubernetes and Docker.
  • Profound experience in SQL and NoSQL databases like MySQL, MongoDB, DynamoDB
  • Expertise in migrating legacy SQLs into DBT tool.

Overview

17
17
years of professional experience

Work History

Principal Data Engineer

Project44
03.2023 - Current
  • Formulated and established robust analytics for shipment tracking and carrier data quality analysis.
  • Built a Data Platform Tool to ingest data from various sources.
  • Lead the transition from batch analytics to real-time analytics, enhancing data processing speed and accuracy.
  • Designed and implemented scalable data pipelines using Apache Kafka and Apache Spark Structured Streaming for real-time data processing.
  • Migrated Shipment Tracking Analytics from distributed models to unified models.
  • Optimized Snowflake ELT pipelines, achieving a 50% reduction in data processing costs.
  • Migrated legacy Snowflake SQLs into DBT tool.
  • Added data quality checks to ensure data reliability.

Project44's Data Analytics Platform enables customers to view real-time and historical shipment performance, select the optimal logistics provider, and allows carriers to monitor shipment health metrics and track process improvements.

Lead Data Engineer

Lowes
08.2021 - 02.2023
  • Leading multiple teams in Data Platform
  • Built Data Ingestion Platform, self service tool to ingest data from various sources
  • Implemented best practices and standards across the Data team
  • Coordinate across retail supply chain, sales, inventory functional teams of data scientists, analysts, product managers, engineers and setup data pipelines for their business operations.
  • Designed and developed RDBMS, Streaming , File ingestion connectors
  • Built data transformation accelerators
  • Designed robust architecture to support HA, Scalability and Data Quality.
  • Collaborate with other teams to incorporate their use cases.
  • Collaborate on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability
  • Integrated Great Expectation tool to ensure pipelines Data Quality
  • Coaching and mentoring engineers on best-in-class data engineering technologies


Data Ingestion Platform tool is self service tool used by lowe's organization to ingest data from various external sources into the Hadoop and transform as per business usecases.

Users can launch data pipelines in few minutes without knowledge on bigdata skills.


Bigdata Architect

Encora Innovation Labs
10.2018 - 08.2021
  • Migrated existing legacy data pipeline stack from ELK to Hadoop stack
  • Designed real time streaming data pipeline to detect treats and notify users.
  • Designing and architecting large scale distributed data management and analytics applications
  • Closely worked with devops team to build Scalable , Highly Available and Robust infra setup on AWS cloud platform.
  • Created data pipeline to process and store massive amount of data in data lake.
  • Developed test automation framework using Robotframework for complete data pipeline quality.
  • Perform proof-of-concept (POC) on new technologies and approaches.

Bigdata Architect

DevOn Software solutions
07.2015 - 09.2018
  • Designed Modules: Data Batch Processing, Data Analytics, Code development, Unit Testing, bug fixes
  • Lead Team members in development activities and work individually on product backlog.
  • Collaborates with product owner and stakeholders in order to understand business needs and provide technical solutions.
  • Make sure the right modelling is being done, to know that qualities like performance are going to be met.
  • Guide in preparation of technical design documents. Ensure that documentation is in place and maintained
  • Provide guidance to the dev team on the best possible implementation/planning during execution of a task.

Tech Lead

Tech4sys Software Solutions
11.2007 - 06.2015

Education

Computer Science

Gokaraju Rangaraju Institute of Engineering
Hyderabad, India
04.2006

Skills

  • Scala, Java, Python
  • Spark, Sqoop, Scala, Kafka, Map Reduce, HDFS, Hive, HBase, Airflow
  • Snowflake, DBT
  • MySQL, PostGreSQL, Teradata, Oracle
  • Kubernetes, Docker
  • Cassandra, Mongo DB, Dynamo DB
  • EMR, Kinesis, S3, Cloud watch, Lambda, Dynamo DB, RDS, Athena
  • Jenkins, Terraform, AWS code build & pipeline, AWS Cloudformation

Certificates

AWS Certified Solutions Architect – Associate (Credential ID: EPSEHZMCN244QB39)

Professional Scrum Master I

Timeline

Principal Data Engineer

Project44
03.2023 - Current

Lead Data Engineer

Lowes
08.2021 - 02.2023

Bigdata Architect

Encora Innovation Labs
10.2018 - 08.2021

Bigdata Architect

DevOn Software solutions
07.2015 - 09.2018

Tech Lead

Tech4sys Software Solutions
11.2007 - 06.2015

Computer Science

Gokaraju Rangaraju Institute of Engineering
Aravind Kumar Anugula