Summary
Overview
Work History
Education
Skills
Certification
Timeline
Receptionist
Venu Katragadda

Venu Katragadda

Technical Lead
Hyderabad

Summary

  • Cloudera Spark , databricks spark and AWS certified professional.
  • Working as a technical Lead to design, development and Support different big data applications by using Apache Spark, Hadoop and AWS.
  • Having 13+ years of experience in IT, include 8+ years’ experience in Apache Spark and 4 years experience in snowflake to implement Big Data Applications
  • Involved in developing Big Data applications by using different frameworks like Hadoop, Hive, Sqoop, Oozie, Spark, Flink, Phoenix, Cassandra, Kafka and Nifi
  • Having extensive experience on Spark SQL, Spark streaming include tune-up of the Spark applications using Scala and Python (Pyspark)
  • Having experience in all stages of the project including requirements gathering, designing & documenting architecture, development, performance optimization, cleaning, and reporting. Experienced in Snowflake.
  • Good experience in Snowflake optimization and cost saving techniques, TimeTravel, Snowflake Optimization techniques.
  • Strong experience on AWS-EMR, Spark Installation, HDFS & Map-reduce Architecture. Along with that having a good knowledge on Spark, Scala and Hadoop distributions like Hortonworks, Apache Hadoop, Cloudera, Azure Talend and Qubole.
  • Strong experience in All Hadoop and Spark ecosystems include Hive, Shell Script, Sqoop, Airflow, Nifi, Kafka, Oozie, Cassandra, Spark SQL, Spark Streaming and Flink.
  • Working on bigdata administration activities like Cloudera, Hortonworks, Apache spark installation as well. Experienced in all major clouds like AWS, Azure and google cloud.
  • Experienced to optimize performance in Snowflake (ETL) projects. ▪ Eager to implement Artificial intelligence projects, having POC level experience in Machine learning (Scikit-Learn), Deep learning (TensorFlow, Keras, PyTorch), Computer Vision (Yolo, OpenCV)
  • Experienced in DevOps tools like Kubernetes, Jenkins, docker, GitHub ▪ Highly experienced to implement Apache Superset to create Dashboards and visualize data from spark and other JDBC urls.

Overview

12
12
years of professional experience
5
5
years of post-secondary education
3
3
Certifications

Work History

Technical Lead

Signin Soft Pvt Ltd
Hyderabad
2023.04 - Current

Professional Experience:

  • Provide expert technical leadership to a team of [X] software developers, ensuring the successful delivery of complex software projects on time and within budget.
  • Collaborate with architects and senior developers to design robust software architectures and select appropriate technologies.
  • Conduct thorough code reviews, ensuring adherence to coding standards and best practices while maintaining a high level of code quality and consistency.
  • Create and maintain comprehensive technical documentation, including design documents and API documentation.
  • Assist in project planning, task estimation, and resource allocation, working closely with project managers to align technical requirements with project goals.
  • Proactively troubleshoot technical issues and lead problem-solving efforts to maintain project momentum and resolve complex challenges.
  • Mentor and coach junior team members, fostering their professional growth and development in software engineering.
  • Stay current with industry trends and emerging technologies, evaluating their applicability to the organization and projects.
  • Facilitate effective communication and collaboration between development teams and various project stakeholders.
  • Oversee code maintenance, including version control best practices, code repository management, and the integration and deployment process.
  • Identify and manage technical risks, developing contingency plans to mitigate potential challenges.
  • Focus on system scalability, performance optimization, and security to ensure the software meets quality and security standards.
  • Continuously advocate for process improvement and implement best practices to enhance efficiency and product quality.

Technical Lead

Lnt Infotech
Chennai
2017.10 - 2022.07

Roles & responsibilities:

Requirement Analysis: Demonstrated a keen understanding of project requirements, ensuring a solid foundation for subsequent tasks.

  • Data Import and Transformation: Effectively imported data from Abinitio, maintaining data integrity and consistency, and then meticulously stored it in Hive based on defined business logic.
  • Job Scheduling and Automation: Implemented efficient data processing by scheduling PySpark jobs using Oozie and creating shell scripts for automation, significantly reducing processing time and manual intervention.
  • Performance Optimization: Proactively enhanced query performance and maintained data accuracy by skillfully utilizing PySpark, optimizing code to meet and exceed performance targets.
  • Performance Analysis: Conducted thorough performance analysis using the Spark web user interface, identifying bottlenecks, and proposing optimizations for improved processing speed.
  • Alternate Code Development: Innovatively crafted alternative code solutions to further boost Spark performance, driving significant performance gains and ensuring efficient data processing.
  • Code Validation and Testing: Demonstrated a strong commitment to quality assurance by rigorously validating and testing the code to ensure robustness, data consistency, and accuracy.
  • Client Collaboration: Actively participated in multivendor discussions with the client throughout the engagement, ensuring seamless collaboration and alignment with project goals.

Pyspark & Snowflake Realtime Data Integration

Everside - EDW Implementation
Chennai
2019.06 - 2022.06

Objective: Showcase your expertise in optimizing data processing by migrating a client project from a costly Snowpipe-based setup to a more cost-effective and efficient solution using Kafka, PySpark, and Snowflake.

Project Details:

Benefits:

  • Demonstrates your ability to optimize data processing workflows for cost savings and efficiency.
  • Highlights your expertise in real-time data integration and data enrichment.
  • Showcases your proficiency in working with cloud data warehousing solutions like Snowflake.
  • Provides evidence of your skill in managing data streams and batch data

Data Migration to Snowflake Data Warehouse

Keurig Dr Pepper
Chennai
2019.07 - 2021.06

Objective: Led a mission-critical data migration project at Keurig Dr Pepper, involving the transition from a Hive/Azure-based data solution to the Snowflake Data Warehouse platform, optimizing data management and query performance for streamlined business intelligence and reporting.

Project Details:

  • Client Overview: Managed data migration for Keurig Dr Pepper, a leading North American coffee and beverage company, formed through the merger of Keurig Green Mountain and Dr Pepper Snapple Group, with market leadership in various beverage categories.

Role & Contribution:

  • Snowflake Solutions Implementation: Spearheaded the implementation of data solutions leveraging Snowflake Data Warehouse, transforming the client's data infrastructure.
  • ETL Pipeline Development: Developed robust ETL pipelines both into and out of the Snowflake Data Warehouse, utilizing a combination of Informatica Cloud and Snowflake's SnowSQL. This process ensured seamless data integration and transfer.
  • SQL Query Development: Proficiently wrote SQL queries against Snowflake, facilitating data retrieval and manipulation for business intelligence and reporting purposes.
  • Testing and Documentation: Rigorously tested and meticulously documented all implementations, ensuring clear communication of requirements, implementation details, and test conditions for the benefit of the team.
  • Data Warehouse Support: Provided ongoing support for Data Warehouse issues, including data load problems and transformation translation challenges. Successfully translated requirements for BI and reporting into database design and reporting design solutions.
  • Data Transformation: Demonstrated a keen understanding of data transformation and translation requirements, selecting appropriate tools and techniques to achieve the desired results efficiently.
  • Database Management: Took the initiative in creating Snowflake databases, schemas, tables, and views, enabling structured and efficient data storage.
  • Component Migration: Led the migration of components from Hive to Snowflake, ensuring a seamless transition to the new platform.
  • Data Loading: Managed data loading processes from source systems, including flat files and Hive tables, to Snowflake using SnowSQL commands. This improved data accessibility and reliability.
  • Query Performance Optimization: Employed various techniques to optimize query performance within Snowflake, enhancing data retrieval efficiency.
  • Resource Management: Utilized the Snowflake Resource Monitor to efficiently manage database objects, ensuring consistent performance and resource allocation.
  • User Access Control: Created Snowflake users and effectively managed access permissions, enhancing data security and control.
  • Historical Data Migration: Developed a strategy for the one-time historical data migration process before the production go-live, ensuring data continuity and completeness.
  • Documentation Preparation: Prepared comprehensive documentation for User Acceptance Testing (UAT) and production migration, facilitating seamless project transitions and ongoing management.

Benefits:

  • Showcases expertise in data migration, data warehousing, and SQL query development.
  • Highlights successful transition to Snowflake Data Warehouse for improved data management and reporting capabilities.
  • Emphasizes proficiency in ETL pipeline development, data optimization, and database management.
  • Demonstrates dedication to rigorous testing, documentation, and end-to-end project support

Streaming Data Analysis &Health Status Management

Capgemini Pvt Ltd, Collabera
bangalore
2017.01 - 2017.07
  • Client Overview: Comcast/Xfinity is a prominent organization handling streaming data from various devices, analyzing router health status, and enhancing data processing for improved network performance.

Role & Contribution:

  • Kafka Data Ingestion: Developed Kafka producer code to efficiently gather streaming data from diverse devices, ensuring real-time data ingestion.
  • Graph-Frames Analysis: Analyzed the relationships between multiple devices using Graph-Frames, contributing to a comprehensive understanding of the network topology.
  • Error Identification: Leveraged Spark and Scala to write code that identified and resolved errors in devices based on business requirements. This process streamlined problem identification and resolution.
  • Spark Optimization: Optimized Spark application jobs, enhancing data read and write efficiency between Oracle and HBase/Phoenix. Achieved resource optimization, resulting in performance improvements.
  • Debugging and Error Resolution: Proficiently debugged Spark jobs, identified and resolved errors, and effectively utilized Spark's web UI for troubleshooting and real-time monitoring.
  • Quality Assurance: Conducted rigorous testing of Spark jobs to verify expected output accuracy and ensure data integrity.
  • Py-Spark Scripting: Developed Py-Spark scripts and designed User-Defined Functions (UDFs) to obtain desired results, enriching data analysis and management.
  • Mentorship: Demonstrated strong leadership by training and mentoring junior team members, fostering their professional growth and ensuring team alignment with project objectives.
  • Client Communication: Maintained regular and effective communication with the client to align project goals and deliver expected outputs, ensuring client satisfaction.

Automation: Successfully automated the entire workflow using shell scripts, resulting in streamlined and efficient data processing.

Benefits:

  • Showcases expertise in streaming data analysis, problem resolution, and performance optimization using Spark and Scala.
  • Highlights proficiency in Graph-Frames and error identification for improved network performance.
  • Emphasizes resource optimization, debugging skills, and testing for quality assurance.
  • Demonstrates leadership through team mentorship and effective client communication.

Data Migrate From Onpremis to Cloud Using Pyspark

Brillio
Bangalore
2015.09 - 2016.06
  • BMSGCM is a health care domain, Get the RDBMS data (Oracle/MySQL) and SFTP data, store in AWS-S3
  • Import and Export the data from OraOop (through Sqoop) and store in S3
  • Apply Transformation rules on the top of different datasets finally store in the desired output (CSV to JSON)
  • Schedule the tasks in AWS-pipes, Scalup automatically based on data
  • Finally store the desired data in Dynamodb, Red
  • Shift and S3 in desired format
  • Create the dashboards, reports using Splunk
  • Role &
  • Contribution For injection data from Oracle, Mysql to S3 which can be queried using hive and spark SQL tables
  • Worked on Sqoop jobs for ingesting data from MySql to Amazon S3 Created hive external tables for querying the data Use Spark Dataframe APIs to inject Oracle data to S3 and stored in Redshift
  • Write a script to get RDBMS data to Redshift
  • Process the datasets and apply different transformation rules on the top of different datasets
  • Process the complex/nested Json and Csv data using Dataframe API
  • Automatically scale-up the EMR Instances based on the data
  • Apply Transformation rules on the top of Datagrams
  • Run and schedule the Spark script in EMR Pipes
  • Process Hive, csv, json, oracle data at a time (POC)
  • Validate and debug the script between source and destination
  • Validate the source and final output data
  • Test the data using Dataset API instead of RDD Debug & test the process is reaching Client's expectations or not
  • Query execution is trigger
  • Improve the process timing
  • Based on new spark versions, applying different optimization transformation rules
  • Debug the script to minimize the shuffling data
  • Technology &Tools
  • SparkSQL, Streaming, Sqoop, AWS (EMR,Elastic Search, S3, EMR, DynamoDB Pipes, Redshift)


Digital Marketing Analyst

Material soft solutions, Venu Katragadda
Hyderabad
2012.04 - 2014.12

Objective: Led a multifaceted project encompassing digital marketing strategies, SEO analysis, PHP development, and web design to enhance online presence and user experience for a client.

Project Details:

  • Client Overview: Collaborated with a diverse client in healthcare industry to improve their online presence, traffic, and user engagement.

Role & Contribution:

  • Digital Marketing Strategies: Developed and executed comprehensive digital marketing strategies, including SEO, content marketing, email campaigns, and social media marketing to drive organic and paid traffic.
  • SEO Analysis: Conducted in-depth SEO analysis, including keyword research, on-page optimization, backlink building, and competitor analysis to improve search engine rankings and visibility.
  • Web Design: Utilized web design expertise to revamp the client's website, focusing on user experience, aesthetics, and mobile responsiveness. Designed user-friendly and visually appealing layouts.
  • PHP Development: Collaborated with the development team to implement custom PHP features and functionalities, such as interactive forms, e-commerce integration, and content management systems, enhancing the website's interactivity.
  • Content Creation: Produced high-quality, SEO-optimized content, including blog posts, articles, and landing pages, to engage visitors and boost search engine rankings.
  • Traffic Analytics: Leveraged Google Analytics and other analytics tools to monitor website traffic, user behavior, and conversion rates. Applied insights to refine marketing strategies and design improvements.
  • PPC Campaigns: Managed pay-per-click (PPC) advertising campaigns on platforms like Google Ads, ensuring targeted ad placements and optimizing ad spend for maximum ROI.
  • A/B Testing: Conducted A/B tests on web design elements, landing pages, and ad copy to refine website performance and conversion rates.
  • E-commerce Optimization: Enhanced the client's e-commerce platform with features like product recommendations, secure payment gateways, and simplified checkout processes to improve user experience and drive sales.
  • Customer Engagement: Implemented strategies to engage with customers, such as live chat support, email newsletters, and user feedback forms.
  • Responsive Design: Ensured the website was fully responsive, providing a seamless user experience on various devices and screen sizes.

Education

Master of Science - Master of Computer Application

Sriram College of Arts And Science
University Of Madras
2006.07 - 2011.11

Skills

    Pyspark

undefined

Certification

Databricks spark certified professional

Timeline

Technical Lead

Signin Soft Pvt Ltd
2023.04 - Current

Databricks spark certified professional

2022-08

AWS Solution architect professional certification

2021-01

Cloudera spark certified professional

2020-01

Data Migration to Snowflake Data Warehouse

Keurig Dr Pepper
2019.07 - 2021.06

Pyspark & Snowflake Realtime Data Integration

Everside - EDW Implementation
2019.06 - 2022.06

Technical Lead

Lnt Infotech
2017.10 - 2022.07

Streaming Data Analysis &Health Status Management

Capgemini Pvt Ltd, Collabera
2017.01 - 2017.07

Data Migrate From Onpremis to Cloud Using Pyspark

Brillio
2015.09 - 2016.06

Digital Marketing Analyst

Material soft solutions, Venu Katragadda
2012.04 - 2014.12

Master of Science - Master of Computer Application

Sriram College of Arts And Science
2006.07 - 2011.11
Venu KatragaddaTechnical Lead