Summary
Overview
Work History
Education
Skills
Certification
Profile Overview
Projects Handled
Timeline
Generic

Hemanth Reddy | Data Engineer

Bangalore

Summary

A motivated individual with in-depth knowledge of languages and development tools, seeking a position in a growth-oriented company where I can use my skills to the advantage of the company while having the scope to develop my own skills. Mainly aspiring to work in data and cloud environments.

Overview

3
3
years of professional experience
1
1
Certification

Work History

Data Engineer

BLACKBUCK INSIGHTS
Bengaluru
03.2021 - Current
  • Design and Development of data pipelines on business requirement
  • Providing Complex Data transformations using PySpark for Data Processing
  • Maintains the reliability and performance of the Data Infrastructure
  • Own the delivery with collaboration across multiple teams
  • Responsible for the Expanding and Optimizing Data pipeline architecture and Data flow
  • Own Technical Documentation
  • Identifying Inconsistencies within the Data
  • Continuous Integration framework using GitHub, BitBucket
  • Continuous delivery framework using Jenkins on multiple environments
  • Providing BigQuery tables as per business need.

Education

Post Graduate Diploma in Big Data Analytics -

Center for Development of Advanced Computing(CDAC-KP)
01.2021

Bachelors in Engineering in Computer Science Engineering -

Sir C R Reddy College of Engineering (Andhra university)
01.2019

Skills

  • PySpark, Python, SQL, Machine Learning, Java
  • Postgres
  • Jenkins, Dremio, DBeaver, Jira, Airflow
  • Amazon Web Services(AWS), Google Cloud Platform(GCP)

Certification

  • Certified AWS Developer Associate Certification(AWS-DAC)
  • Certified AWS Data Analytics Specialty Certification(AWS-DAS)
  • Certified DataBricks PySpark Developer
  • Certified Astronomer Certification DAG Authoring for Apache Airflow
  • Certified Astronomer Certification for Apache Airflow fundamentals

Profile Overview

  • Data Pipeline development using Airflow and AWS Services like S3, EMR, Lambda.
  • Working on Common Code Repositories like GitHub and BitBucket.
  • Awareness of API services in GCP.
  • Exposure of Cloud Technologies like AWS and GCP.
  • Working experience on data file formats like Parquet, csv, psv, tsv, avro, txt.
  • Knowledge of Software Development Processes & Models: Agile, Waterfall.
  • Experience of batch processing of the files into/from s3 and processing them with EMR.
  • Experience in maintaining CI environments with build automation tools like Jenkins.
  • Experience in developing SQL queries and dashboards in BigQuery and Looker as per business requirement.
  • Developed lambda functionality to auto terminate EMR Clusters and Cloud Functions to Terminate Data Proc Clusters.
  • Experience in working with PostGres database.
  • Excellent analytical and programming abilities in using technology to create flexible and maintainable solutions for complex development problems.

Projects Handled

  • Data Services | AVANT, DS Team Member, AVANT is a Fin-Tech Organization which deals with credit cards and loans in US, due to huge volume of its customer base they required a data processing system to process and analyze their customers and provide them efficient services. for building up that particular data processing system, we have built a platform that includes Airflow as Orchestration tool, AWS as Cloud storage and processing machines like EMR and Kafka, other tool like Dbeaver for Database, Dremio for Data Orchestration from S3, GitHub for Common Code Repository.
  • Data Services | AEO, Core Engineering Team Member, American Eagle Outfitters(AEO) which is a leading global specialty retailer offering high-quality, on-trend clothing, accessories and personal care products. It has a requirement to migrate in-house data and processing machines to Cloud. We have used Google Cloud Platform for the requirement and migrated the data to GCP cloud storage buckets and used GCP composer for the data processing. Along with that for the client need we have used GCP BigQuery for the Orchestration of the Raw/Processed Data. We have used tools like Terraform for Deployments and BitBucket for Common Code Repository.

Timeline

Data Engineer

BLACKBUCK INSIGHTS
03.2021 - Current

Post Graduate Diploma in Big Data Analytics -

Center for Development of Advanced Computing(CDAC-KP)

Bachelors in Engineering in Computer Science Engineering -

Sir C R Reddy College of Engineering (Andhra university)
Hemanth Reddy | Data Engineer