
Data Engineer with 6 years of experience designing, building, and optimizing scalable data pipelines, real-time streaming systems, and analytics platforms. Strong background in SQL, Kafka, Redis, Docker, Kubernetes, and cloud-native architectures. Proven ability to work across data ingestion, transformation, storage, and API layers to support analytics, reporting, and AI-driven systems. Adept at collaborating with product, analytics, and backend teams to deliver reliable, high-performance data solutions.
Data Engineering: Data Pipelines, ETL / ELT, Batch & Real-Time Processing
Streaming & Messaging: Apache Kafka, Event-Driven Architectures
Distributed Processing: Apache Spark, PySpark
Data Warehousing: Amazon Redshift, BigQuery, Snowflake
Orchestration: Apache Airflow, Dagster
Data Lake / Lakehouse: S3-based Data Lakes, Lakehouse Architecture Concepts
Databases: MySQL, MongoDB, Redis
Programming: Java, Python (Data Processing & Automation)
APIs & Integration: REST APIs, Data Services
Containerization & DevOps: Docker, Kubernetes
Cloud & Storage: AWS (S3), Object Storage, IAM
Version Control: Git
Performance & Reliability: Query Optimization, Monitoring, Debugging