Summary
Overview
Work History
Education
Skills
Certification
Timeline
Generic

Aman Kumar

Kolkata

Summary

Results-driven Data Engineer with nearly 5 years of experience in designing and implementing scalable ETL/ELT pipelines, data models, and cloud data solutions. Expertise includes Snowflake (SnowPro Core certified), DBT, Python, and big data frameworks such as Spark and Hadoop. Proven ability to build automated data pipelines, optimize queries, and enable analytics for large-scale enterprises. Skilled in collaborating with cross-functional teams to deliver high-quality, business-critical data products that drive strategic decision-making.

Overview

3
3
years of professional experience
1
1
Certification

Work History

Specialist Programmer L2 (Data Engineer)

Infosys Ltd.
06.2022 - Current
  • Project: Data Center Intelligence (DCI)
  • Designed and developed ETL pipelines to ingest structured & semi-structured data from APIs, RDBMS, AWS S3 into Snowflake.
  • Implemented DBT models for data transformations (SCD Type 2, incremental loads, aggregations), improving pipeline modularity and reusability.
  • Integrated DBT runs with Airflow for end-to-end automation and lineage tracking.
  • Optimized Snowflake queries and warehouse performance, reducing query execution time by 30% through clustering keys and caching.
  • Built a generic Python framework for API data extraction with pagination, token handling, multithreading, and logging, reducing development effort by 40%.
  • Automated CI/CD pipelines using Python and Shell scripts, lowering manual effort by 80% and improving reliability.
  • Collaborated with BI teams to build data models supporting real-time operational dashboards.
  • Implemented CDC-based incremental pipelines into Snowflake, reducing data latency and improving reporting efficiency.
  • Project: MOAE – Marketing Operational Analytical Environment
  • Migrated legacy ETL workflows from RDBMS/HDFS to Scality S3 using PySpark and Dremio.
  • Converted Sqoop & Hive queries into Spark SQL & Python-based data pipelines.
  • Developed Python utilities for Hive table creation and handling file formats (Parquet, JSON, CSV).
  • Automated Jenkins cleanup jobs via Tivoli scheduler to improve reliability.
  • Built virtual datasets in Dremio for analytics and reporting, enabling marketing teams with faster insights.
  • Collaborated with cross-functional teams to streamline processes and enhance project delivery efficiency.
  • Analyzed business requirements to develop effective solutions aligned with organizational goals.

Education

B.Tech. - Computer Science and Engineering

MAKAUT
West Bengal
07-2020

Skills

  • Programming & Scripting: Python, SQL, Shell, PySpark, Pandas
  • Data Warehousing & Modeling: Snowflake, Star/Snowflake schema, SCD, CDC
  • Transformation & Orchestration: dbt, Apache Airflow, ETL/ELT design, Jenkins
  • Cloud Platforms: AWS (S3, Glue, Lambda, EC2, IAM, CloudWatch)
  • Big Data & Storage: Hadoop, Scality S3, Parquet/CSV/JSON handling
  • Version Control & DevOps: Git, Docker, Kubernetes, Ansible, CI/CD automation
  • Visualization & Documentation: Dremio, Tableau/Power BI (exposure), Miro, Quip

Certification

  • SnowPro Core Certified – Snowflake Inc.
  • Infosys Certified Python Developer
  • Infosys Certified Big Data Developer
  • AWS Generative AI Learning Plan – Developers

Timeline

Specialist Programmer L2 (Data Engineer)

Infosys Ltd.
06.2022 - Current

B.Tech. - Computer Science and Engineering

MAKAUT
Aman Kumar