Summary
Overview
Work History
Education
Skills
Photo
Timeline
Generic

Jeevani Potturu

Hyderabad

Summary

Accomplished Data Engineer specializing in Scala Spark and PySpark, refined at Unilever. Demonstrated success in architecting scalable data pipelines that improved data accuracy and accessibility, facilitating strategic decision-making. Skilled in collaborating with cross-functional teams to deliver robust data solutions aligned with business objectives.

Overview

8
8
years of professional experience

Work History

Data Engineer

Unilever
11.2023 - Current
  • Company Overview: Unilever is a global leader in consumer goods, offering products in nutrition, hygiene, and personal care, with a strong commitment to sustainability and innovation. A multinational company with a diverse portfolio of trusted brands including Dove, Lipton, Knorr, etc.. Unilever operates in over 190 countries, serving billions of consumers daily.
  • Designed and maintained scalable data pipelines to process secondary sales data at Unilever, enabling real-time insights for sales and distribution teams across multiple markets.
  • Engineered data solutions for Unilever’s secondary sales systems, improving data accuracy and accessibility for business stakeholders, and supporting strategic decision-making in retail and distribution.
  • Responsible for handling the Data Landscape of the Project.
  • Worked on End-to-End Model Building and EDA for Mexico.
  • Involved in End-to-End Model Building of Ever billed and Width Pack KPI’s of Indonesia.
  • Worked on End-to-End Data Analysis of Cost to Serve Module for Vietnam.
  • Worked on collection and Validation of the data.
  • Conducted detailed requirements analysis and involved in Data Preparation and Transformations using Scala Spark and Pyspark using Azure Databricks.
  • Involved in working on Data Analysis like Clustering, Feature creation and Model Building.
  • Involved in generating the IQ KPI Performance score for the build Model.
  • Collaborated with cross functional stakeholders including data scientists and business analysts to define data requirements and deliver robust data solutions.
  • Worked closely with the clients to understand their data needs, providing customised solutions.
  • Responsible for generating and documenting the monthly Reports related to the Project.
  • Unilever is a global leader in consumer goods, offering products in nutrition, hygiene, and personal care, with a strong commitment to sustainability and innovation. A multinational company with a diverse portfolio of trusted brands including Dove, Lipton, Knorr, etc.. Unilever operates in over 190 countries, serving billions of consumers daily.
  • Technologies: Scala Spark, PySpark, Azure Databricks

Data Engineer

Cisco
11.2020 - 09.2023
  • Company Overview: Cisco wanted to build a platform to provide a fully governed central analytical solution built on modern scalable architecture (Azure cloud and Data Bricks); provision data from the systems of records, business critical insights to Cisco Products, Services, Sales, Returns, Market reach and some view over the customers data associated with the company.
  • Responsible for developing and Implementing solutions using PySpark API to load data from different RDBMS Servers into ADLS Platform.
  • Responsible for doing the analysis on Resiliency Data.
  • Worked with the internal and external teams to resolve any issues surrounding the data pipelines.
  • Involved in creating Hive tables, data loading and writing hive queries.
  • Worked on Scheduling the Jobs through Azure Data Bricks Workflows.
  • Build processes supporting data transformation, data structures, metadata, dependency and workload management.
  • Involved in working on Data Analysis, Data quality and data profiling for handling business that helped the business team.
  • Understand and build re-usable data assets or features to enable downstream data science models to read and use.
  • Cisco wanted to build a platform to provide a fully governed central analytical solution built on modern scalable architecture (Azure cloud and Data Bricks); provision data from the systems of records, business critical insights to Cisco Products, Services, Sales, Returns, Market reach and some view over the customers data associated with the company.
  • Technologies: Spark-SQL, PySpark, Jira, Hive, Cloudera, Agile Methodology, MySQL

Data Engineer

Infosys
07.2019 - 10.2020
  • Company Overview: Employees learning R & A is one of the centralized source systems which contains organization level trainings, budget and employee related data.
  • Develop, test and maintain the data infrastructure, the data ingestion pipeline, data store and data processing on premise.
  • Working with business analysts in gathering business requirements, obtaining clarifications and approvals during functional analysis, data validations and integration testing.
  • Worked on different file formats like Text file, Sequence file, Avro, Json, Parquet, Orc, Csv, Xml, and Custom delimited file formats.
  • Imported data from MySQL to DBFS using Sqoop.
  • Involved in creating Hive tables, loading data and writing hive queries using Spark-SQL.
  • Involved in partitioning and joining hive table for hive query optimization.
  • Worked on Data Lake and Staging Area where data will be staged into AVRO file formats.
  • Writing Test cases for the developed code.
  • Importing the huge amounts of retail data into HDFS from various sources and processing it in CLOUDERA.
  • Performed Import and Export of data into HDFS and Hive using Sqoop and managed data within the environment.
  • Involved in creating Hive tables, data loading and writing hive queries.
  • Was responsible for Optimizing Hive queries that helped in saving Cost to the project.
  • Managed Hive Tables and created child tables based on partitions.
  • Used spark transformations while moving the data to STAGING layer.
  • Employees learning R & A is one of the centralized source systems which contains organization level trainings, budget and employee related data.
  • Technologies used: Spark-SQL, PySpark, Azure ADLS, Jira, Hive, Azure Data Bricks, Hive, Agile Methodology, MySQL, AWS S3, CLOUDERA CDH

Test Analyst

11.2017 - 06.2019
  • Worked as a Test Analyst on the Light Application and MDM Application.
  • Responsible for Testing the My Deal Manager Application.
  • Reviewing the User Stories Related to the enhancements of the MDM Application.
  • Preparing the Scenarios from the User Stories.
  • Reviewing the Scenarios with the Business Team.
  • Preparing the Test Cases from the Scenarios.
  • Uploading the Test cases in Rally.
  • Executing the test cases.
  • Updating the Results in Rally.
  • Raising the defects for the test cases failed during the execution in Rally.
  • Retesting the raised Defects.
  • Handling the UAT during execution.

Education

B. Tech -

Jawaharlal Nehru Technological University
Anantapur
01.2013

Skills

  • Python
  • MySQL
  • Hadoop
  • Sqoop
  • Hive
  • HDFS
  • Spark-SQL
  • Databricks
  • DBFS
  • Azure
  • Eclipse
  • Git
  • Maven
  • JIRA
  • Jupiter Notebook
  • PyCharm
  • PySpark
  • Scala Spark

Photo

30KB, JPEG, in formals with light background

Timeline

Data Engineer

Unilever
11.2023 - Current

Data Engineer

Cisco
11.2020 - 09.2023

Data Engineer

Infosys
07.2019 - 10.2020

Test Analyst

11.2017 - 06.2019

B. Tech -

Jawaharlal Nehru Technological University
Jeevani Potturu