Over all 9+ years of professional experience and 5+ years of relevant expertise in Big Data technologies, including Hadoop, Spark, Hive, delivering scalable data pipelines, ETL solutions, and analytics workflows on CDP and AWS. Skilled in AWS services (S3, IAM), SQL query optimization, Hive (Partitioning, Bucketing, Aggregation), and Spark architecture (Spark Core, Spark SQL, DataFrames). Proven track record of implementing and developing Big Data analytics initiatives and managing large-scale data processing solutions. Proficient in the Hadoop ecosystem (HDFS, Hive, PySpark), Linux, and CI/CD pipelines using Jenkins, Autosys, RLM. Experienced with development tools and version control: Intellij, PyCharm, GitHub, Bitbucket. Strong expertise in production support, defect resolution, RCA, release management, and code reviews. Analytical and fast learner with excellent problem-solving and debugging skills, adaptable to emerging technologies. Strong problem solving and technical skills coupled with clear decision-making and ability to learn quickly and adapt to new technologies. Adept in collaboration, mentorship, and cross-functional teamwork, fostering a culture of continuous improvement. Experience in developing the projects using Agile Methodology. Experienced in tracking incidents via JIRA and ensuring smooth project delivery through Agile best practices.
Client : Citi Bank
Project: The AML Markets platform handles large-scale trade and transaction data from global markets to detect potential suspicious activities, insider trading, and money laundering patterns. It ingests high-volume data from various source systems (trade bookings, pricing systems, reference data), applies complex business rules, and generates alerts for investigation by compliance teams.
Roles:
Technology & Tools: PySpark, HIVE, HDFS, Shell scripts ,VS code ,RLM, Jenkins, Notebook ++,Vim,CDP, AWS (S3,IAM), Linux
Client : DBS Bank
DBS is a leading financial services group in Asia with a presence in 18 markets. As a part of the team needs to maintain, process huge amount of data as part of day-to-day operations. As a Big Data Engineer, worked with Risk Management team to insert the processed data in to Hive Table using the Spark Code
Roles:
Technology & Tools: PYSPARK, HIVE, HDFS, SQOOP & CDH, Bitbucket
Client: Logitech
Project: Global Data Engineering & Analytics Platform
Roles:
Technology & Tools: PyCharm, Visual Studio code, Python, Pyspark,Hive,Shell Scripting, Jupyter notebook, Notepad++, Vim
Client: Adobe
Project: A call center customer experience management solution (Hendrix)
Roles:
Skills: IntelliJ, Core Java, Html, XML and CSS, Selenium WebDriver, SQL Server, Jira, Maven
Company: Prime Focus technologies
Project: Clear Media ERP (Star Sports)
Roles:
Project Specific Skills: Core Java, Html, XML and CSS, Selenium WebDriver, JavaScript, SQL Server, Jira, Maven