Summary
Overview
Work History
Education
Skills
Relevant Experience
Domains
Overall Experience
Timeline
Generic

Ranjeet Kumar

Data Engineer Professional
Pune

Summary

Total 9 Year of experience in IT industry where 7 year of experience across Big Data Hadoop, Datawarehouse, Datalake, Data Modeling PySpark, Spark, Hive, HDFS, Python, Shell script and Linux. Ability to quickly pick up unfamiliar technologies, which in turn allows swiftly diagnosing problems and keeping everything running smoothly. Troubleshooting data related and other technical issues on production.

Overview

10
10
years of professional experience

Work History

Lead Technology

Synechron
Pune
02.2023 - Current

Project and Team : Synechron Innovation

  • Deployed and gained hands-on expertise with the Open Metadata framework, leveraging its features for data quality, data insights, and data governance.
  • Developed conceptual, logical, and physical data models to support business intelligence, reporting, and data warehousing requirements.
  • Ensured data quality and consistency by implementing normalization, indexing, and dimensional modeling (star/snowflake schemas) for complex datasets.
  • Implemented data masking on Hive tables using Apache Ranger to enhance data security and compliance.
  • Troubleshot and resolved infrastructure-level issues across Hive, Airflow, PySpark environments to ensure optimal system performance and reliability.

Senior Software Engineer

Clairvoyant India Private Limited
Pune
05.2018 - 01.2023

Project: Cross Property DataLake
Client: PayPal,
Role: Senior Data Engineer

  • We had different types of PayPal subsidiary like Xoom, Venmo, Hyperwallet, Swift. Managed and implemented the end-to-end data pipeline, data quality check to run the data ingestion process smooth.
  • Involved in on-premises to GCP data migration activity using Big Query, Data proc and GCS.
  • Handled and Processed TB’s of data from different sources like aws s3, redshift for PayPal subsidiary Xoom, Venmo, Swift, Hyperwallet, and iZattle.
  • Handling ETL Pipeline with data ingestion into data warehouse and responsible for data quality checks.
  • Involved in data analysis for missing records on source and target tables. and performed backfilling and history load.
  • Managing cluster level issues if anything is failing due to infra level issue like hive issue or spark session issue.
  • Responsible for end to end deploying new jobs if any requirement comes from business.
  • Handled cluster to cluster replication process.
  • Loading and transformed large sets of data.
  • Involved in job management and scheduling using UC4 and airflow.
  • Automate the data purging for hive tables and NFS with retention of period using shell script.
  • Analysis if any jobs are long running or stuck.
  • Resolve the issue to fulfill the requirement within SLA.
  • Environment: Hadoop Multi node cluster, Hive, HDFS, MySQL, Spark, Linux, Airflow, UC4

Software Engineer

AMDOCS INDIA PVT.LTD. (Parent Company: VSOFT)
Gurgaon
11.2017 - 04.2018

Project: Telecom
Client: Bharti Airtel,
Role: Software Engineer


• Generating reports for the BI team by exporting data into HDFS and Hive using Sqoop.
• Importing, exporting, and transforming large volumes of structured, semi-structured, and unstructured data in HDFS and Hive.
• Monitoring, managing, and analyzing Hadoop process and file to ensure smooth job execution and data integrity.
• Managing incoming data from diverse sources and automating data pipelines to improve operational efficiency.
• Handling job scheduling and resource management using Fair Scheduler for optimized cluster performance.
• Coordinating Hadoop cluster operations and performing troubleshooting and maintenance tasks.
• Working extensively with Linux environments, Apache Tomcat, and SQL database servers for operational support.
• Automating routine tasks, handling support tickets, and resolving issues within defined SLA timelines to ensure service reliability and client satisfaction.

Software Engineer

Mcarbon Tech Innovation Pvt. Ltd.
Noida
09.2016 - 09.2017

Project: Telecom
Client: Airtel, Aircel, Tata Docomo
Role: Support Engineer

Description: I was responsible for providing end-to-end platform support, data management, and operational excellence across Linux-based environments. The work focused on ensuring data availability, system reliability, and seamless client coordination.

• Loading day end process file means incremental data into oracle database.
• Monitoring application and it’s log files from Linux servers.
• Automating backup scheduling and various operational jobs through custom scripting.
• Providing platform support for Linux-based servers, ensuring high system availability.
• Troubleshooting service and application issues promptly to maintain service continuity.
• Monitoring critical services and server health to prevent potential outages and disruptions.
• Coordinating actively with clients for planned activities, ensuring clear communication and minimal business impact.
• Preparing and maintaining comprehensive documentation, user manuals, End-of-Day (EOD) procedures, and Root Cause Analysis (RCA) reports.
• Managing change processes through effective change management and change coordination practices.
• Handling service outages with a structured approach to minimize downtime and impact.
• Ensuring timely response and resolution of incidents within the agreed SLA, maintaining high service quality standards

Software Engineer

IBM INDIA Pvt.Ltd. (Parent Company: Net Connect)
Noida
10.2015 - 08.2016

Project: Telecom (Bharti Airtel)

  • Operations-Service Provisioning and Telecom Business Operation (TBO- Prepaid) for the telecommunication services (Provisioning/De provisioning, VAS–START/STOP, Blackberry services Activation/Deactivation, Rules for Benefits passing to the customers etc.).
  • Health Check monitoring of jobs and Linux different zones server.
  • Handling in the form of SR(service request) & IN(Incident).
  • Perform activity like Server bounce, DB bounce, PIBS Cold backup activity etc.
  • Worked on activities such as password change, database management and testing on behalf of business.
  • Loading CDR’s for the purpose of billing & reporting.
  • Looking after provisioning & de-provisioning related issues on behalf of database.
  • Analyzing application problems, recommending and developing solutions.
  • Provided support during various phases of testing as well as production support.
  • Monitoring database and linux server performance.

Education

Master In Computer Application - Computer Applications Development

Indira Gandhi National Open University
New Delhi
08.2013

Skills

  • Hive

  • PySpark

  • Spark

  • SQL

  • Python

  • Shell Script

  • Linux

  • ADLS

  • Hadoop

  • Databricks

  • Azure Cloud

  • GCP Migration

DataLake

Relevant Experience

6 Years

Domains

  • Banking
  • Finance

Overall Experience

9 Years

Timeline

Lead Technology

Synechron
02.2023 - Current

Senior Software Engineer

Clairvoyant India Private Limited
05.2018 - 01.2023

Software Engineer

AMDOCS INDIA PVT.LTD. (Parent Company: VSOFT)
11.2017 - 04.2018

Software Engineer

Mcarbon Tech Innovation Pvt. Ltd.
09.2016 - 09.2017

Software Engineer

IBM INDIA Pvt.Ltd. (Parent Company: Net Connect)
10.2015 - 08.2016

Master In Computer Application - Computer Applications Development

Indira Gandhi National Open University
Ranjeet KumarData Engineer Professional