Summary
Overview
Work History
Education
Skills
Websites
Languages
Career Experience
Professional Summary
Technology Proficiency
Personal Information
Disclaimer
Timeline
AdministrativeAssistant

Somnath Dadasaheb Wandhekar

Pune,Maharashtra

Summary

  • To secure a challenging position in a reputable organization to expand my learnings knowledge and skills ,while making significant contribution to the success of the company

  • Overall 3.8 years of experience in database management and Development.
  • Extensive experience which includes Hadoop Eco System and Big Data analytics.
  • Have hands on experience in writing Pyspark jobs in Python , Hive , Sqoop, NoSql Databases such as Hbase , Spark and hands on installing, configuring admin Hadoop cluster.
  • Excellent analytical, problem solving, communication and interpersonal skills with ability to interact with individuals and can work as a part of a team as well as independently.
  • Proficient in analyzing and translating business requirement to functional/technical requirements.
  • Good experience SQL such as Oracle , Postgre , MySQL.
  • Committed team player with good logical and problem solving skills, willing to relocate and ability to quickly adapt to new environments & technologies.
  • Having experience in cloud (AWS- iam,s3,emr,Athena,glue).
  • Have worked on huge dataset of structured, non-structured data.
  • Understanding of Git branching and merging strategies
  • Hands on with pycharm & Databricks
  • Basic knowledge of azure cloud technologies such as azure Databricks.
  • Having the basic knowledge of aws cloud technologies

Overview

3
3
years of professional experience

Work History

Data Engineer

Facile services
Pune
05.2022 - Current
  • Importing data from PostgreSQL and Oracle To Hbase using Phoenix integration
  • Have hands on experience in writing Pyspark jobs in Python , Hive , Sqoop, NoSql Databases such as Hbase , Spark and hands on installing, configuring admin Hadoop cluster.
  • Data Ingestion in Hadoop Data lake.
  • Data validation using Quality check.
  • Job execution monitoring, debug, bug fixing.
  • Written spark jobs as per business requirement.
  • Deduplication and reconciliation of huge dataset
  • Understanding of how to deploy application using CI/CD pipeline
  • Used tools ETL, Hands-on experience in deploying applications using CI/CD pipelines (Git, Jenkins).
  • Have Worked on performance optimization of code level and execution level
  • Worked on data processing, data analysis, data modeling, automation, tuning, and data ingestion using PySpark, SQL, and Python.

Education

BE - Computer Engineering

University of Pune

Skills

  • Windows
  • Ubuntu
  • Redhat
  • Python
  • Hadoop
  • HDFS
  • YARN
  • Map Reduce
  • Sqoop
  • Hive
  • Hbase
  • Spark
  • Zookeeper
  • Pycharm
  • MySQL
  • Postgre
  • Oracle
  • AWS
  • Cron tab
  • AirFlow

Languages

English
Hindi
Marathi

Career Experience

Facile services, Pune, India, Data Engineer, Credit card data analysis, 2023-09-01, Present, Importing data from PostgreSQL and Oracle to Hbase using Phoenix integration, Writing Pyspark jobs in Python, Hive, Sqoop, NoSql Databases such as Hbase, Spark, Installing, configuring admin Hadoop cluster, Data ingestion in Hadoop Data lake, Data validation using Quality check, Job execution monitoring, debug, bug fixing, Written spark jobs as per business requirement, Deduplication and reconciliation of huge dataset, Understanding of how to deploy application using CI/CD pipeline, Used tools ETL, deploying applications using CI/CD pipelines (Git, Jenkins), Worked on performance optimization of code level and execution level, Data processing, data analysis, data modeling, automation, tuning, and data ingestion using PySpark, SQL, and Python, Data ingestion in data warehouse, 2022-05-01, 2023-08-31, Understand the upstream source nature and work with business cases, Written sqoop logic to extract records from downstream source, Work closely with the business and analytics team in gathering the system requirements, Export and Import batch data and delta into HDFS, HBase and Hive using Sqoop, Creating Hive tables, loading with data and writing Hive queries, Analyzed Data in data warehouse, Strong knowledge of Version control tool - GIT, Expert or advance knowledge of Unix/Linux operating system, Reconcile the imported or processed data as per business requirement, Used tools Hive, Kafka, GITLAB, YARN, Jira, Scrum

Professional Summary

Hadoop Eco System, Big Data analytics, Pyspark jobs in Python, Hive, Sqoop, NoSql Databases such as Hbase, Spark, Hadoop cluster installation and configuration, SQL (Oracle, Postgre, MySQL), Cloud (AWS - iam, s3, emr, Athena, glue), Git branching and merging strategies, Pycharm, Databricks, Azure cloud technologies, Excellent analytical, problem solving, communication and interpersonal skills, Ability to interact with individuals and work as part of a team or independently, Committed team player with good logical and problem solving skills, Willing to relocate and ability to quickly adapt to new environments & technologies

Technology Proficiency

Windows, Ubuntu, Redhat, Python, Hadoop, HDFS, YARN, Map Reduce, Sqoop, Hive, Hbase, Spark, Zookeeper, Pycharm, MySQL, Postgre, Oracle, AWS (Iam, Emr, Ec2, S3, Rds, Iam, Athena), Cron tab, AirFlow

Personal Information

  • Title: Data Engineer
  • Date of Birth: 10 Nov 1997

Disclaimer

I hereby declare that the above given information is true and to the best of my knowledge.

Timeline

Data Engineer

Facile services
05.2022 - Current

BE - Computer Engineering

University of Pune
Somnath Dadasaheb Wandhekar