Summary
Overview
Work History
Education
Skills
Hobbies and Interests
Languages
Personal Information
Languages
Professional Experience
Timeline
Generic
Md Rahmatullah

Md Rahmatullah

Pune

Summary

Possess 3.5+ years of professional experience in IT.Experienced Senior AWS Data Engineer and Data Architect with a strong background in building and managing scalable cloud data platforms. Skilled in Python, PySpark, Pandas, and REST APIs, with expertise in large-scale data migration, data cleansing, transformation, and warehouse management. Demonstrated success in developing end-to-end ETL pipelines for efficient data processing and enabling analytics-driven business insights.

Proficient in leveraging AWS services such as S3, SageMaker, EC2,Cloudwatch,Lambda,IAM and Glue to set up Cloud Data Lakes, Data Warehouses, and support AI/ML model deployments. Played a key role in data platform migration and analytics consolidation projects for global clients such as HDBFS and Aliaxis, enabling seamless reporting and insights generation.

Hands-on experience with Data Science tools like Jupyter Notebook and AWS SageMaker for developing models and cloud-based analytics solutions. Adept at Data Science as a Service (DSaaS), identifying relevant use cases to drive business value through advanced data analytics.

Experienced in developing SaaS platforms like RESENSE for retail and consumer brands, with strong skills in object-oriented programming for web applications. Expertise in collaborating with cross-functional teams, peer code reviews, and version control using GitLab.

Proven ability to understand business requirements, build custom data solutions, and create efficient business logic using Python. Passionate about consolidating data, developing cloud capabilities, and delivering scalable, data-driven insights for enterprise-level transformations.

Overview

3
3
years of professional experience

Work History

Data Lake Implementation

HDB Financial Services
12.2022 - Current
  • Company Overview: HDB Financial Services (HDBFS) is a non-banking financial company (NBFC) headquartered in Mumbai, India
  • It was founded in 2007 and is a subsidiary of HDFC Bank Limited, one of the largest private sector banks in India
  • HDBFS provides a wide range of financial services to individuals and small and medium-sized enterprises (SMEs)
  • Its offerings include loans, insurance, credit cards, savings accounts, and other banking and financial products
  • The company has a presence in over 1,000 cities in India and has a customer base of over 10 million
  • Written Python scripts in AWS EC2 Instance virtual machine to Extract large volume data(10 billions records from one table) in chunk-size of 200000 from Oracle database tables(220 tables ,source FINNLOS,FINLMS,FINCOLL) and directly uploaded to AWS s3 bucket in raw-zone
  • Written Python scripts cleansed and formatting large volume data using pandas
  • Written python scripts to parse nested dense complex flies from lentra source and converted into csv files using python and pandas and loaded into s3 bucket
  • Written python scripts to download zipped file from sftp and unzipped these files into s3 folder
  • Written python scripts for convert dbf(karvy,cams) and bson files into csv files
  • Written Python scripts in EC2 instance to fetch data from salesforce soql using salesforce rest api and bulk api and directly uploaded into aws s3 bucket in raw-zone
  • Written logic for full load incremental load using python
  • Written and automate Python scripts to move sap file s3 bucket inbound to s3 bucket raw-zone using boto3 python library
  • Created tables in amazon redshifts and written store procedure and populated historical and incremental data into the tables
  • Created tables in amazon redshift database for different purpose like to maintain job detail,for master file and for incremental details
  • Merge files into single file from s3 bucket using aws glue pyspark
  • Automate python scripts using Task Scheduler
  • Used Postman tools for testing sfdc restapi and bulkapi using salesforce object query language
  • HDB Financial Services (HDBFS) is a non-banking financial company (NBFC) headquartered in Mumbai, India
  • It was founded in 2007 and is a subsidiary of HDFC Bank Limited, one of the largest private sector banks in India
  • HDBFS provides a wide range of financial services to individuals and small and medium-sized enterprises (SMEs)
  • Its offerings include loans, insurance, credit cards, savings accounts, and other banking and financial products
  • The company has a presence in over 1,000 cities in India and has a customer base of over 10 million
  • Technologies: SSH Linux, Arcon, Python, Oracle, AWS Cloud, Amazon Redshift, Data Brew, Salesforce Rest API and BulkApi

ADP-Phase1

Sapient Wealth
06.2022 - 12.2022
  • Company Overview: Sapient Wealth is a financial services company founded in 2009
  • The company offers a range of services, including investment management, financial planning, tax planning, insurance, and retirement planning
  • Written Class based Python scripts in AWS EC2 Instance virtual machine to download karvy and cams dbf files from s3 bucket and converted these dbf files into csv files and uploaded into s3 bucket in respective folder according to the business requirements
  • Written python scripts for calculating capital gain and xirr for karvy and cams files
  • Written Python scripts for calculating transaction aggregation and balance for karvy and cams source files and loaded into postgresql database
  • Written Python scripts cleansed and formatting large volume data
  • Written Various Python scripts according to business requirements and loading these datas into postgresql
  • Written logic for full load and incremental load using python
  • Written and automate Python scripts to move dbf files s3 bucket inbound to archive folder using boto3 python library
  • Created tables in postgresql database for different purpose like to maintain job detail,for master file and for incremental details
  • Written and transform all python codes into pyspark
  • Loaded karvy and cams files into postgresql tables using pyspark using aws glue
  • Written pyspark code in aws glue to load historical and incremental data into various tables
  • Automate python scripts using Task Sheduler
  • Automate aws glue service using triggers
  • Sapient Wealth is a financial services company founded in 2009
  • The company offers a range of services, including investment management, financial planning, tax planning, insurance, and retirement planning
  • Technologies: Python, Postgresql, AWS Cloud, MySQL, Pyspark, Pandas

Data Lake Implementation

Ashirvad Pipes
02.2022 - 07.2022
  • Company Overview: Ashirvad Pipes was founded in 1998 & became a part of Aliaxis Group in 2018
  • It has embarked on a Digital Transformation journey to ensure that their vision to double the revenue and wish to build a Data Road map
  • It uses Qlik Sense for reporting
  • Written Python scripts in AWS EC2 Instance virtual machine to Extract large volume data in chunksize from Mysql db( soursesystem Sparsh,Exellence Club,Dealer Portal and WST) and directly uploaded to AWS s3 bucket in rawzone
  • Written Python scripts cleansed and formatting large volume data
  • Developed product recommendation model using Collaborative filtering
  • Written Python scripts in EC2 instance to fetch data from salesforce soql using salesforce rest api and bulk api and directly uploaded into aws s3 bucket in rawzone
  • Written logic for full load incremental load using python
  • Written and automate Python scripts to move sap file s3 bucket inbound to s3 bucket rawzone using boto3 python library
  • Created tables in snowflake database for different purpose like to maintain job detail,for master file and for incremental details
  • Merge files into single file from s3 bucket using aws glue pyspark
  • Written Aws Glue code using pyspark to move file and formatting from rawzone to process zone
  • Written AWS Glue Code using pyspark to create table in snowflake prod ods from s3 bucket files
  • Automate python scripts using Task Sheduler
  • Automate aws glue service using triggers
  • Automate pipleline from data extraction to Snowflake Prod ods
  • Used Postman tools for testing sfdc restapi and bulkpapi using salesforce object query language
  • Ashirvad Pipes was founded in 1998 & became a part of Aliaxis Group in 2018
  • It has embarked on a Digital Transformation journey to ensure that their vision to double the revenue and wish to build a Data Road map
  • It uses Qlik Sense for reporting
  • Technologies: Python, Snowflake Cloud, AWS Cloud, Salesforce Rest API and BulkApi, MySQL, SQL Server, AWS Sagemaker & Machine Learning

POC

Nihilent Ltd
11.2021 - 02.2022
  • Company Overview: RESENSE – a Cloud hosted a Service (SaaS) platform offers range of services covering across business functions for retail and consumer brands
  • Written two Restful API in Flask for Try Resense Now and Book a Consult Now
  • Integrated communication channel (email servcer side)
  • Implemented machine learning model
  • Implemented various service to send email notification
  • Analyses and working on Model classes
  • Writing the business logic with the help of Python & Flask
  • Involving coding and review of the System
  • Used Postman for testing the functionality of the Restful API
  • RESENSE – a Cloud hosted a Service (SaaS) platform offers range of services covering across business functions for retail and consumer brands
  • Technologies: Python, Flask, Rest API, Oracle, Angular, AI, Machine Learning

Data Lake Implementation

Tata Power Delhi Distribution Ltd
06.2021 - 10.2021
  • Company Overview: Tata Power Delhi Distribution Ltd (TPDDL, hereafter) is the country’ most successful and innovative private power distribution utility company
  • Written python scripts to fetch data from different soursesystem
  • Cleansed data using python and pandas library
  • Tata Power Delhi Distribution Ltd (TPDDL, hereafter) is the country’ most successful and innovative private power distribution utility company
  • Technologies: Python, Aws, SQL Server, Sap Hana, Hadoop, Oracle, Web Api, & Machine Learning

Education

BE - E&TC

SKNCOE
Pune
01.2014

Skills

  • Python
  • SQL
  • Pandas
  • Pyspark
  • Flask
  • AWS
  • AWS Lambda
  • AWS Glue
  • AWS S3
  • Mysql
  • Snowflake
  • Redshift
  • Oracle
  • Pycharm IDE
  • VS Code
  • Jupyter Notebook
  • DBeaver
  • Windows
  • SSH-LINUX
  • Mac
  • Jupyter Notebook
  • DBeaver
  • Windows
  • SSH-LINUX
  • Mac
  • Jupyter Notebook
  • DBeaver
  • Windows
  • SSH-LINUX
  • Mac
  • Jupyter Notebook
  • DBeaver
  • Windows
  • SSH-LINUX
  • Mac
  • Jupyter Notebook
  • DBeaver
  • Windows
  • SSH-LINUX
  • Mac

Hobbies and Interests

Reading Books

Languages

  • English
  • Hindi
  • Urdu

Personal Information

  • Father's Name: Md Farid Rahmani
  • Nationality: Indian

Languages

English
First Language
Urdu
Elementary (A2)
A2

Professional Experience

Senior Software Developer – Data Engineering and Cloud Solutions
Nihilent Ltd, Pune | June 15, 2021 – Present

  • Lead data migration, cleansing, and transformation projects for large datasets, ensuring seamless integration into data warehouses and cloud platforms.
  • Develop and maintain end-to-end ETL pipelines using Python, PySpark, and AWS Glue to enhance data accessibility and analytics capabilities.
  • Set up Cloud Data Lakes and Data Warehouses on AWS to support reporting and AI/ML models for clients.

Timeline

Data Lake Implementation

HDB Financial Services
12.2022 - Current

ADP-Phase1

Sapient Wealth
06.2022 - 12.2022

Data Lake Implementation

Ashirvad Pipes
02.2022 - 07.2022

POC

Nihilent Ltd
11.2021 - 02.2022

Data Lake Implementation

Tata Power Delhi Distribution Ltd
06.2021 - 10.2021

BE - E&TC

SKNCOE
Md Rahmatullah