To work in a challenging environment and effectively utilize my skills to contribute towards the goals of the organization, besides being resourceful and innovative.
2+ years of experience in Bigdata and Hadoop ecosystems like HDFS, MapReduce, Yarn, Sqoop, Hive and Oozie.
Load and transform large sets of structured, semi-structured and unstructured data from Relational Database Systems to HDFS and vice-versa using Sqoop tool.
Having work experience in SQL. Experience in AWS Glue, AWS Athena, AWS S3.
Resourceful,and committed to continually advancing my knowledge of relevant technologies.
Experienced and dedicated Data Analyst with several years of experience identifying efficiencies and problem areas within data streams, while communicating needs for projects.
Adept at receiving and monitoring data from multiple data streams.
Overview
3
3
years of professional experience
Work History
Data Analyst
GENPACT
Hyderabad
11.2021 - Current
Involved in extracting data from various data sources into Hadoop HDFS.This included data from AWS.
Maintain the day-to-day data in Hive Datawarehouse without lagging from different source systems.
Writing the Sqoop Jobs to transfer data from SQL Server and Oracle Databases to HDFS;
Scheduling the Sqoop jobs using Crontab.
To Check the SQL Database for monitoring daily ingestion jobs are successful if not debugging the issue.
Worked on creating Hive managed and external tables based on the requirement,
Configured Oozie workflow to run multiple Hive and spark jobs which run independently with time and data availability.
Configuring Oozie to check the dependency data before the workflow get started.
Creating the tables in Athena.
Education
Skills
BIG DATA ECOSYSTEMS: Hadoop, Spark, Hive, HBase, Yarn