Overall 8+ years of experience in IT Industry. Around 3 years of Hands-on experience as Android Developer. Around 5+ years of Hands-on experience as DATA Engineer. Responsibilities Handled: Having experience in data ingestion, storage, querying, processing and analysis of data in Google Cloud Ecosystem using GCS, Big Query, Airflow, DataProc, PUBSUB. Having experience in data ingestion, storage, querying, processing and analysis of data in Hadoop Ecosystem using Map Reduce, HDFS, Hive, Sqoop, and Spark. Capable of processing large sets of structural and semi-structural and un-structured and supporting system application architecture. Very good understanding on developing ETL pipeline for customers using Spark, Scala, Hive and various BIGDATA tools. Managed data coming from different sources and involved in HDFS. Having hands on experience in HDFS Data Flow to process the data in batch mode. Hands-on experience in writing Sqoop jobs, to implement the import and export data from HDFS to RDBMS and vice versa. Worked in Agile methodologies. Developed a robust data-pipeline to cleanse, filter, aggregate, normalize, and de-normalize the data using Apache Spark. Working knowledge in different file formats like Parquet, ORC, XML, JSON, CSV files processing in Hive and spark. Hands-on experience in writing hive queries for data analysis to meet the business requirements with the improve performance by implementing the Partitioning, Bucketing concepts. Hands-on experience in creating Datasets and Data Frames in spark SQL. Having strong knowledge in MR1/MR2 (YARN) Architectures. Having good knowledge in Kafka. Quick learner and self-starter to go ahead with any new technology.
Technologies: GCP Services, Language, Databases, Bigdata, Hadoop Distribution, Scripting language, Version control tools, Scheduling tools, IDE’s
undefined