Around 10 years of technical expertise in various domains that includes financial and telecom with hands-on experience in Big Data Analytics design and development Around 4+ years of relevant experience in Big Data Analytics and data manipulation using Hadoop Ecosystem tools such as MapReduce, HDFS, Yarn/ MRv2, Hive, HBase, Spark, Kafka, Sqoop, Oozie, Avro, Kerberos and Spark Integration with Cassandra and Zookeeper along with ADF, EventHub, Synapse, Databricks & Azure Function. Rich experience in designing and developing applications in Apache Spark, PySpark, Scala, Java, Kafka, Python, and Hive with Hadoop Ecosystem along with NRT process. Strong experience on Hadoop distributions with Cloudera. Hands-on experience in working on Spark, PySpark, RDD, DataFrame, and Dataset API for processing unstructured and structured data Efficient in writing live real-time processing and core jobs using Spark Structured Streaming with Kafka as a data pipeline system Well versed in writing multiple jobs using Spark and Hive for data extraction, transformation, and aggregation from multiple file formats including Parquet, Avro, XML, JSON, CSV, and OrcFile and other compressed file formats codecs such as GZIP, Snappy.