Big Data Engineer around 5 years of experience in designing, developing, and optimizing high-performance data pipelines and architectures. Experienced data engineer with a specialization in Big Data processing and distributed computing. Proficient in designing and optimizing ETL/ELT workflows for seamless data integration using Python, SQL, Apache Spark and Azure, AWS
Project Name: Emergency Respond Data Warehouse Analysis
Project Name — Telecom Infra Data Analysis
Big Data Technologies: Hadoop, Apache Spark, Python, SOL, Hive, PySpark, Yarn, Hdfs, Map Reduce
Cloud Technologies: Azure(ADF, Synapse, Azure Databricks), AWS(S3,RDS,ATHENA,EMR)
Databases and Tools: Oracle, MySQL, NoSQL, PostgreSQL, HBase
Languages: Python, SQL
Scheduling: Apache Airflow, Cron Tab
Data Modeling: Data Warehousing , Data Lake
Data Pipelines: ETL, ELT, Streaming, Batch Processing, Data Strategy, Data Source, Production Issues, Data Transformation
Data Governance: Data Quality, Data Security, Data Lineage , Metadata Management
Version Control: Git, Jira , Agile, Scrum
Platforms: Windows, Linux