Summary
Overview
Work History
Education
Skills
Timeline
Generic

MANO PRIYA R

Google Certified Professional Data Engineer
Chennai

Summary

Results-driven data engineering professional with a robust background in designing and maintaining scalable data systems. Expertise in implementing ETL data pipelines leveraging GCP services ensuring efficient data management and processing. Proficient in crafting PySpark jobs and Linux Shell scripts for ETL processes, data analysis, and automation, leading to significant operational improvements. Demonstrated success in performance tuning of ETL jobs while collaborating closely with clients to gather requirements and design tailored systems, showcasing a strong commitment to delivering end-to-end solutions from product development to delivery.

Overview

11
11
years of professional experience
3
3
Languages

Work History

Data Engineer

Altimetrik
06.2023 - Current

Ford Motors Global Data Insights and Analytics

  • With predominant experience in pyspark, python and GCP services, worked in the Global Data insights and Analytics in understanding and formulating business problem statements and converting business problem statement into data science problems.
  • We receive the A to Z of a Vehicle like Parts information, software and hardware parts, updates, services and history. Worked on constructing meaningful pipelines, constructing daily/weekly batches, constructing models of pipelines, Visualizing and sending to business to aid in Business decision making.

Roles & Responsibilities:

  • Data Ingestion and end to end Pipeline designing from Storage and management to transformation and Processing.
  • Architected scalable systems and data infrastructure supporting batch/real-time processing of billions of records, automating cloud services, CI/CD pipelines, and observability.
  • Collaborated cross-functionally with data scientists, engineers, and product teams to gather requirements, solve business problems, and deliver actionable insights through analytics and dashboards.
  • Applied a product-focused and growth mindset to identify opportunities, optimize KPIs, and drive revenue impact via strategic data mining, reporting, and stakeholder presentations.
  • Collaborated on ETL (Extract, Transform, Load) tasks, maintaining data integrity and verifying pipeline stability.
  • Automated routine tasks using Python scripts, increasing team productivity and reducing manual errors.
  • Evaluated various tools, technologies, and best practices for potential adoption in the company's data engineering processes.

Data Engineer

Cognizant Technology Solutions
02.2020 - 05.2023

Business Enhancement&Customer Assistance Platform for Insurance Project

  • Project focuses on analyzing policies that are being sold in different states of US on premiums, network of hospitals , other companies policies, accident statistics, people into different risk categories to improve Customer Acquisition and retention, risk assessment, predictive modeling, Personalized Service and Pricing.
  • Majorly worked on collecting data from sources with unstructured/ structured data, constructing end to end pipelines where data is extracted, scrubbed, analyzed, transformed and staged for consumption by Strategy& Modelling Team.

Roles & Responsibilities:

  • Designed dataflow and implemented end to end pipeline.
  • Designed Pyspark RDD/SQL jobs to scrub/clean huge datafile and load them into hive table
  • Designed Pyspark job to replace legacy sqoop rdbms ingestion pipeline
  • Designed data validation framework having standard validation checks like Recon, counts
  • Developed atomic scripts for scheduling daily/weekly jobs.
  • Worked on Agile based environment using jira life cycle management tool.

Performance Tuning:

  • Derived and altered partitions based on the input parameters.
  • Applied bucketing techniques on Join queries involving large tables.
  • Solved skewing problems using clustering.
  • Worked on JSON(File format) and File Types like AVRO, ORC, Parquet.
  • Changed from dynamic partitioning to static partitioning wherever there was scope to reduce sorting overhead.
  • Redesigned resource allocation on spark jobs.
  • Applied optimization techniques like Push down optimization, Predicate Pushdown and partition pruning wherever required .
  • Resolved Skewness in Spark programs by applying techniques like repartition and shuffle partition configurations.

Associate

Cognizant Technology Solutions
09.2014 - 01.2020

Bankers Interactive Claim Processing System (BICPS) is Claim adjudication and Processing application. It processes Long Term care (LTC), Medicare Supplement, Specified Disease claims and Major medical claims.

Roles & Responsibilities:

  • Worked on Understanding the business and building various batch, online modules.
  • Developed Cobol programs from scratch based on program specifications as per applicable language standards.
  • Involved in App based Pricing.
  • Worked on a windows server migration project end to end on the Dev and Prod servers.

Education

Bachelor of Technology - Information Technology

Anna University
TamilNadu
04.2001 -

Skills

Google Cloud Services: DataProc, DataFlow, BigQuery, Pub / Sub, Cloud Logging

Timeline

Data Engineer

Altimetrik
06.2023 - Current

Data Engineer

Cognizant Technology Solutions
02.2020 - 05.2023

Associate

Cognizant Technology Solutions
09.2014 - 01.2020

Bachelor of Technology - Information Technology

Anna University
04.2001 -
MANO PRIYA RGoogle Certified Professional Data Engineer