Summary
Overview
Work History
Education
Skills
Certification
Websites
Timeline
Generic

SAGAR DNYANE

Summary

Looking for a challenging role in a reputable organization to utilize my technical skills for the growth of the organization as well as to enhance my knowledge about new and emerging trends in the IT sector. I am looking forward to work in an organization, which provides the environment for growth with competent and strong leaders in Python/Aws developer ,Data Scientist. CAREER HIGHLIGHTS Overall 7 years of experience in Data Scientist/Data Engineer and Techniques, covering a wide range of understanding business objectives and developing models that help to achieve them. I have experience in working as Data Scientist and Data Engineer. Experience in working with AWS Lambda, AWS Connect, Amazon Lex, AWS CLI, AWS SDK with Python-Boto3, Serverless , sagemaker,EMR, Kinesis, Glue, S3, RDS and other core AWS services. Worked extensively on S3, CloudWatch, CloudFront. Experienced in developing the data pipelines and managing the ETL end to end with AWS glue. Working experience in Database design, Data Modelling, Performance tuning and query optimization with large volume of multi-dimensional data. Having experience in analyzing the ML algorithms that could be used to solve a given problem. Having experience in data analytics, analyse large datasets, Filter and clean data, Exploratory Data Analysis Identify, analyze, and interpret trends or patterns in complex data sets. Exploring and visualizing data to gain an understanding and identifying differences in data distribution. Develop and implement the databases, data collection systems and data analytics. Identify, analyze and interpret trends or patters in complex datasets. etc . Data analytics Techniques and its Techniques like Numpy, Pandas, Matplotlib, Seaborn, PCA for Data Analysis etc. Machine Learning Techniques and its Techniques like Regression, Classification etc. Involved in Classical Optimization (gradient descent) Having strong skills on Python, AWS services,Tableau, Databricks, MongoDB, Mysql . Highly competent Data Engineer with background in designing, testing, and maintaining data management systems. Possess strong skills in database design and data mining, coupled with adeptness at using machine learning to improve business decision making. Previous work resulted in optimizing data retrieval processes and improving system efficiency.

Overview

8
8
years of professional experience
1
1
Certification

Work History

Data Engineer

Netlink Software Pvt Ltd
Pune
11.2024 - Current

Project: AbbVie Migraine BOTOX Commercial Model.

Client Name: ZS Associates India Pvt. Ltd.

Role: AWS Data Engineer.

  • Collaborated with stakeholders to define business requirements for data analytics projects.
  • Performed exploratory data analysis using Pandas, NumPy, and openpyxl libraries in Python.
  • Created ETL processes using Python scripts to move data from various sources into the target databases on AWS Redshift.
  • Developed and maintained data pipelines to ingest, store, process, and analyze large datasets in AWS S3 buckets.
  • Implemented automated monitoring of data flows using Cloud Watch and Lambda functions.
  • Collaborated with stakeholders across business units to define requirements for new BI initiatives, leveraging AWS services like Athena, QuickSight, and Glue.

Database Engineer

Sense And Respond software LLP
12.2023 - 07.2024

Project : Renew Financial

Role: Database Engineer (Contract to hire Position)

  • Optimized existing queries to improve query performance by creating indexes on tables.
  • Analyzed user requirements, designed and developed ETL processes to load enterprise data into the Data Warehouse.
  • Developed and implemented data models, database designs, data access and table maintenance codes.
  • Created stored procedures for automating periodic tasks in SQL Server.
  • Collaborated with data scientists and analysts to understand data needs and implement appropriate data models and structures.
  • Configured and maintained cloud-based data infrastructure on platforms like AWS Cloud to enhance data storage and computation capabilities.
  • Conducted data analysis using SQL and Python to derive insights and support decision-making processes.

Sr.Software Engineer (Python/Aws / Data Engineer)

Virtusa consulting services pvt ltd
Pune
04.2021 - 12.2023

Project :RBS

Client : CITI BANK (2021 To till date)

Role:Aws Data Engineer

Description:

Developed security polices and processes.setup and build aws infrastructure various resource,VPC,EC2,S3,IAM,EBS,Security Group,Auto scaling and RDS Cloud Formation JSON template.

Key Responsibilities:

✔ Extensively working with AWS services like Aws glue,S3, EC2, ELB, EBS, Lambda, Auto-Scaling, Route53, CloudFront, IAM, Cloud Watch, and RDS ,Pyspark etc.

✔ Involved in working on Spark SQL Code as an alternative approach for faster data Processing and better Performance. S

✔ Design and Develop ETL Processes in AWS Glue to migrate Campaign data from external sources like S3, ORC/Parquet/Text Files into AWS Redshift.

✔ Manage service account credentials, database credentials, passwords through AWS Secrets Manager.

✔ Imported required data from source to target (RDS) location using Glue job.

✔ Involved in monitoring and handling Glue job.

✔ Writing required logic to load as well as extract necessary data according to client requirements.

✔ Involved in creating an Oracle table loading data, and writing a query that will run internally.

✔ Extracting CSV or XML files according to the customer’s requirement at desired S3 location.

✔ Validating the records once the transformation glue job is completed.

✔ Technologies Skills: AWS, Spark, Python, Oracle

✔ Performed S3 buckets creation, policies on IAM role-based policies, MFA and customizing the JSON template.

✔ Understanding of secure-cloud configuration, Cloud Trail, cloud-security technologies (VPC, Security Groups, etc.) and cloud-permission systems (IAM)

Client : Bancorp Bank

USA Role: Data Scientist

Key Responsibilities:

✔ Developed applications of Machine Learning, Statistical Analysis and Data Visualizations with challenging data Processing problems in banking domain.

✔ Compiled data from various sources public and private databases to perform complex analysis and data manipulation for actionable results.

✔ Applied linear regression, multiple regression, ordinary least square method, mean- variance, theory of large numbers, logistic regression, dummy variable, residuals, Poisson distribution, Bayes, Naive Bayes, fitting function to data with help of Scikit learn, Scipy, NumPy and Pandas module of Python.

✔ Designed and developed Natural Language Processing models for sentiment analysis.

✔ Worked on Natural Language Processing with NLTK module of python for application development for automated customer response.

Resolved complex issues related to software applications quickly and effectively.

Performed unit testing on application components prior to release.

✔ Launched Amazon EC2, EMR Cloud Instances using Amazon Images (Linux/ Ubuntu) an

✔ Configuring launched instances with respect to specific applications.

Environment:Machinelearning,AWS,Spark,Pyspark,NosqlTableau,Python(Scikit-Learn/Scipy/NumPy/Pandas),MySQL, Eclipse, SQL, Linux, SQL connector,Glue,Redshift,Lambda.

Software Engineer (Data Scientist/Data Engineer)

Radical Software
Noida
12.2016 - 04.2021

Project : Smart Transportations System Role: Data Analyst

Description:

In this project basically works for Landon transportation system the main objective of this project to identify various kind of disruption on road according to real traffic this result achieve using EDA technique on the analysis base applying various machine learning algorithms try to find out some prediction.

Key Responsibilities:

✔ Interpret data, analyze results using statistical techniques and provide ongoing reports.

✔ Aims to achieve traffic efficiency by minimizing traffic problem.

✔ Extract real time streaming data using tfl api store in cloud server.

✔ To analyze the current traffic & disruption for specific route recommend possible alternate routes.

✔ Develop and implement databases, data collection systems, data analytics and other strategies that optimize statistical efficiency and quality.

✔ This mended collective data is analyzed further to predict traffic scenario which is available to deliver appropriate information to users.

✔ Identify, analyze, and interpret trends or patterns in complex data sets.

✔ Analysed large datasets, Filter and clean data.

✔ Data Wrangling with Pandas, Exploratory Data Analysis.

Project : Housing Finance company

Role: Data Scientist

Description: Company wants to automate the loan eligibility process (real time) based on customer detail provided while filling online application form. These details are Gender, Marital Status, Education, Number of Dependents, Income, Loan Amount, Credit History and others. To automate this process, they have given a problem to identify the customers segments, those are eligible for loan amount so that they can specifically target these customers. Here they have provided a partial data set.

Key Responsibilities:

✔ Selecting features, building and optimizing classifiers using machine learning techniques

✔ Experience in predictive modelling and predictive software development

✔ Data mining using state-of-the-art methods

✔ Extending client's data with third-party sources of information when needed

✔ Enhancing data collection procedures to include information that is relevant for building analytic systems

✔ Processing, cleansing, and verifying the integrity of data used for analysis

Statistical Method for Data Analysis, Application Of Inferential Statistics, Bayesian Inference, A/B Testing, Frequent Inference and Descriptive statistics.

Education

Bachelor of Engineering -

MGM Jawaharlal Nehru Engineering College (IT)

Skills

  • Experience in predictive modelling and predictive software development
  • Extensively working with AWS services like Aws glue, S3, EC2, ELB, EBS, Lambda, Auto-Scaling, Route53, CloudFront, IAM, Cloud Watch, and RDS etc
  • Created Terraform scripts for EC2 instances, Elastic Load balancers and S3 buckets
  • Extensively working with Docker
  • Skilled in Databricks,Python 38 version (or similar scripting language)
  • Experience in using R, Matlab, or any other statistical software
  • Experience in mentoring junior team members, and guiding them on machine learning and data modelling applications
  • Strong communication and data presentation skills
  • Agile methodology
  • Using with databricks data pipelines and monitor and manage resources in Databricks
  • I have 4years experience pyspark and SQL ,Nosql database
  • Data Warehouse: RedShift, Snowflake
  • Platforms: Linux/Ubuntu, Windows
  • Database: Oracle, MySQL
  • Language & Tools: Python & PyCharm

Certification

B.E(IT) passing year 2016.

PG Data Science and Engineering

Red Hat (RHCA)

AWS Certified Solutions Architect - Associate(Registration Number: 439963964)

Timeline

Data Engineer

Netlink Software Pvt Ltd
11.2024 - Current

Database Engineer

Sense And Respond software LLP
12.2023 - 07.2024

Sr.Software Engineer (Python/Aws / Data Engineer)

Virtusa consulting services pvt ltd
04.2021 - 12.2023

Software Engineer (Data Scientist/Data Engineer)

Radical Software
12.2016 - 04.2021

Bachelor of Engineering -

MGM Jawaharlal Nehru Engineering College (IT)
SAGAR DNYANE