Facile services, Pune, India, Data Engineer, Credit card data analysis, 2023-09-01, Present, Importing data from PostgreSQL and Oracle to Hbase using Phoenix integration, Writing Pyspark jobs in Python, Hive, Sqoop, NoSql Databases such as Hbase, Spark, Installing, configuring admin Hadoop cluster, Data ingestion in Hadoop Data lake, Data validation using Quality check, Job execution monitoring, debug, bug fixing, Written spark jobs as per business requirement, Deduplication and reconciliation of huge dataset, Understanding of how to deploy application using CI/CD pipeline, Used tools ETL, deploying applications using CI/CD pipelines (Git, Jenkins), Worked on performance optimization of code level and execution level, Data processing, data analysis, data modeling, automation, tuning, and data ingestion using PySpark, SQL, and Python, Data ingestion in data warehouse, 2022-05-01, 2023-08-31, Understand the upstream source nature and work with business cases, Written sqoop logic to extract records from downstream source, Work closely with the business and analytics team in gathering the system requirements, Export and Import batch data and delta into HDFS, HBase and Hive using Sqoop, Creating Hive tables, loading with data and writing Hive queries, Analyzed Data in data warehouse, Strong knowledge of Version control tool - GIT, Expert or advance knowledge of Unix/Linux operating system, Reconcile the imported or processed data as per business requirement, Used tools Hive, Kafka, GITLAB, YARN, Jira, Scrum