Summary
Overview
Work History
Education
Skills
Websites
Accomplishments
Personal Information
Languages
Hobbies and Interests
Timeline
Generic

Nagula Moksha

Hyderabad

Summary

Over 3 years of experience as Data Engineer in areas including AWS Cloud Engineering, Data Migration, Batch Scripting with large data sets of structured data source. Experienced in building end to end data solutions using PostgreSQL, Teradata SQL, Linux by leveraging Data Mining. Working experience on various phases like Analysis, Implementation, Testing and Deployment. Experienced in Agile Methodologies, Scrum stories and sprints experience in a SQL based environment. Experienced in Amazon Web Services (AWS) such as AWS Glue, Lambda, S3, RDS, Redshift, Amazon Managed Airflow, Athena, Cloudwatch. Experienced in creating tables, views and schemas within project and have Knowledge in migrating project from one environment to another environment(i.e., Dev to Stage and Stage to Production). Experienced in processing large datasets with PostgreSQL using Teradata SQL. Understanding in scheduling and managing code deployments with DevOps team and maintain Confluence notes. Knowledge of big data technologies like Apache Spark, Hadoop and Gathr. Gained Knowledge in validation and debugging the improper data. Experienced in integration of various relational and non-relational sources such as SQL Server, NoSQL-DynamoDB. Passionate about technology and driven to learn more services related to AWS. Dedicated to delivering extraordinary results that help to growth and success.

Overview

3
3
years of professional experience

Work History

Data Engineer

Mphasis Limited
Chennai
04.2024 - Current

UDHOpsCore Support Project

Description of Project:

  • Efficiently managed and troubleshooted issues within the AWS infrastructure, utilizing the abilities of the AWS Console to ensure optimal operation.
  • Managed system performance and availability through the use of AWS CloudWatch and various monitoring tools.
  • Handled user requests and addressed infrastructure issues through resolution of Jira and Service Now tickets.
  • Monitor and respond to operational emails to ensure timely resolution of incidents.
  • Ensured optimal performance of Airflow DAGs through diligent monitoring and troubleshooting using essential tools like Hubble, Gather, DBeaver, Putty, Load Server, and ABI Server.
  • Managed the development and maintenance of operational documentation and runbooks for incident response and system maintenance.

Software Developer

Mphasis Limited
Chennai
07.2023 - 04.2024

UAInc-Apps-UDH-DR-Mobilize

Description of Project:

  • Copied pre-requisites of airline data from primary AWS region to secondary region for services including S3, RDS, Glue, Redshift, Airflow, Gathr, and ICON.
  • Safeguard data integrity by implementing a backup strategy for disaster recovery.
  • Utilize cloud infrastructure to store and process subsequent sets of data in the secondary region.

Roles and Responsibilities:

  • Validated and replaced hardcoded values in glue scripts, secret_manager, yml, and Bteq files for primary and secondary regions. Ensured DR compliance by creating airflow variables instead of hardcoded values. Verified data in secondary Redshift.

Software Developer

Mphasis Limited
Chennai
08.2021 - 06.2023

UA-RTF-UCLOUD

Description of Project:

  • The aim of this project is to migrate United Airlines' data from Teradata to AWS Cloud, with the primary objective of making AWS the primary data storage for United Airlines.
  • The process involves extracting data in batches and processing it by creating a Direct Acyclic Graph (DAG) in the Managed Apache Airflow Service. Multiple tasks are associated with the Dag, and each task has its own functionality to process data and validate task instances through glue jobs.
  • The data which is extracted from source is loaded into temporary tables though ICON(Ingest, Catalog, Optimize and Notify) process which acts as a gatekeeper for incoming data into data lake and loading to target tables in redshift or RDS Database based on the details mentioned in secret managers.

Roles and Responsibilities :

  • Designed tables in RDBMS (e.g., PostgreSQL) such as RDS and processed data, including importing and exporting databases to AWS Redshift from Aurora and Athena tables
  • Creating and uploading the YAML files related to a particular feed in AWS S3 and validating the DAG in Managed Apache Airflow Service.
  • Handled large datasets of various data types with several kinds of interval columns during the data processing.
  • Developed batch loads using PLSql within the Redshift framework and integrated data from various sources (JDBC, RDBMS, text files, DAT, CSV)
  • Utilized Putty for data extraction and configured Utility for Test Harness to test source and destination data with several checks including alpha numeric and sum check etc
  • Worked with AWS Glue Jobs to monitor Airflow DAG progress and troubleshoot errors
  • Designed and implemented an ingestion template to generate SQL and HQL files for loading preliminary ingestion into an Athena table, and then transferring it to Redshift for comparison.
  • Utilized S3 and several buckets designed to hold files from incremental to backup to post-consumption (RAW bucket) files
  • Analyzed data types compatible with Teradata and addressed AWS Redshift's limitations on supported data types
  • Imported data from Teradata into Athena via SFTP configurations designed in DAG files and loaded it into AWS Athena with compatible data types for Redshift/RDS
  • Imported data from Athena to Redshift/RDS using Airflow and Glue job configurations set up in DAG files
  • Analyzed BTEQ scripts and devised solutions for implementing batch load functions
  • Configured Gathr Pipeline to import data from Teradata to RDS tables, including data type conversion and matching with PySpark scripting
  • Automated data workflows and scripts using Amazon Managed Airflow to ensure scheduled execution in production
  • Understanding in logs monitoring within Airflow and Glue Jobs, utilizing this knowledge for effective debugging and troubleshooting
  • Created and updated the confluence page with complete process and validation of loads
  • Promoted the loads to higher environments by committing in GIT.

Education

Bachelor of Technology in Computer Science, First class -

Marri Laxman Reddy Institute of Technology and Management, Hyderabad
01.2021

Skills

Technical Skills:

  • MySQL
  • AWS(S3, RDS, IAM, AWS Glue, Secret manager, Redshift, Apache Airflow, ETL, Athena)
  • Real-time data processing(Kafka and Kinesis)
  • Data Validation and debugging
  • Data Monitoring & Alerting
  • Apache Spark
  • Linux
  • Python Scripting
  • YAML Configuration

Application Software:

  • SQL Server
  • Jira Software
  • WinSCP(ESFTP Server)

Tools:

  • DBeaver
  • Confluence
  • Putty/ABI Load server
  • ServiceNow
  • Jupyter

Accomplishments

  • KUDOS: Awarded “KUDOS” thrice for completing the critical work load which has 1500+ lines of code within a week. Developed and fixed the issues of another critical load in AWS RDS.
  • Pinnacle: Awarded “Pinacle” once for completing one of the complex load which has 20 tables/Views.
  • Synergic work Force: This is team achievement for migrating the project and received a positive feedback from client.

Personal Information

Title: Software Engineer | Data Engineer | SQL Developer

Languages

  • English
  • Telugu
  • Hindi

Hobbies and Interests

Flexibility and willingness to take on new challenges

Timeline

Data Engineer

Mphasis Limited
04.2024 - Current

Software Developer

Mphasis Limited
07.2023 - 04.2024

Software Developer

Mphasis Limited
08.2021 - 06.2023

Bachelor of Technology in Computer Science, First class -

Marri Laxman Reddy Institute of Technology and Management, Hyderabad
Nagula Moksha