Summary
Work History
Education
Skills
Timeline
Generic

Siva K

HYDERABAD

Summary

· Dedicated Data Engineer with 3 years of experience in Google Cloud Platform (GCP), assisting organizations in handling and migrating on-premise ETLs to GCP utilizing BigQuery.

· Proficient in utilizing GCS, BigQuery, Airflow/Composer, IAM, Cloud Run, Scheduler, and Pub/Sub within the GCP environment.

· Expertise in extracting data from the Landing zone and loading it into target systems.

· Designed and implemented complex workflows using Apache Airflow/Google Cloud Composer to orchestrate and schedule data pipelines effectively.

· Hands-on experience with GCS, BigQuery, SQL, Airflow/Composer, SDK, Cloud Shell, GSUTIL, gcloud function, BQ commands, cloud dataflow, and Pub/Sub.

· Managed Google Cloud Storage (GCS) buckets and objects for efficient data storage and retrieval, with additional proficiency in Java (core concepts) including multithreading and exception handling.

· Extensive experience in developing complex SQL queries to extract, transform, load, and analyze data from various source systems.

· Proficient in monitoring Airflow/Composer DAGs (Directed Acyclic Graphs), defining dependencies, and task sequencing to ensure efficient data processing.

Skilled in Agile methodology, utilizing tools such as JIRA, GitHub, Confluence, and working with various file formats including CSV, JSON, Avro, and Parquet. Strong understanding of BigData Concepts such as Hadoop and PySpark.

Work History

Data Engineer

dafaf
03.2024 - Current

· Worked on Creating DAG's and tasks to deploy and run jobs in the Airflow scheduling tool. Used Source code Management (SCM) as GI Hub.

· Responsible for ingesting the data from one dataset to different datasets through BQ.

· Queries as a part of ELT and ETL process. Responsible for SIT walk through to Business to get Sign-off. Responsible for Generating the Test reports through Python Automation framework.

· Responsible for providing Data Quality Assurance.

· Responsible for analyzing the requirements and designing the requirements page.

· Responsible for creating the GCP buckets, Datasets, Big Query tables in different layers of different BQ projects.

· Created Authorized views in BigQuery to provide data to Down Streams.

· Involved in Development of creating Data Pipeline to read the Data from Kafka and data to Big Query.

Data Engineer

Data
02.2024 - Current

· Working with terms creates various level metric and supporting data pipelines written GCP stack working with large sets of data and solving difficult analytical problem as per the client requirement on Google cloud platform service Such as big query, cloud pub/sub.

· Used cloud shell SDK in GCP to configure the serving data proc, storage, big query Troubleshooting user’s analysis bugs (JIRA AND IRIS TICKETS).

· Crated ETL pipeline using spark to ingest data from multiple sources.

· Responsible for creating the GCP buckets, Datasets, BigQuery tables in different layers of BQ.

· Participation in review Clients calls.

Education

B-TECH(72%) -

BVC Engineering College Amalapuram
Amalapuram
07-2033

Diploma(68%) -

KIET Engineering College

Skills

  • SQL and Databases

Timeline

Data Engineer

dafaf
03.2024 - Current

Data Engineer

Data
02.2024 - Current

B-TECH(72%) -

BVC Engineering College Amalapuram

Diploma(68%) -

KIET Engineering College
Siva K