Dynamic data engineer with a proven track record of working on scalable data pipelines for large datasets. Proficient in Python and SQL programming, optimizing database performance. Experienced in leveraging AWS services for cloud solutions, data transformation, ETL pipelines, and data warehousing. Skilled in analyzing complex data, identifying issues, and drawing sound conclusions. Committed to continuous learning and implementing best practices for innovative solutions. Strong written and oral communication skills, with a proactive approach to contributing beyond basic responsibilities.
Python Programming
Blue Lake, Airline [ANA - All Nippon Airways] - Japan, Migration to New Technology. Blue Lake is a data infrastructure that combines a data lake and data warehouse that makes the most of cloud technology and services while leveraging existing systems to achieve high flexibility, robustness, and economy., Collecting data from different sources depends on business systems and databases and gets cleansed, mapped, and transformed into a quality data set loaded to Redshift based on the schema level., Moreover, push the cleansed data to Targeted Databases., With the help of AWS services, creating each job for transforming data using PySpark in AWS Glue., Calling the glue job using shell scripts, and loading data to Redshift are the main operations present in Bluelake.