Data Science
A Hadoop Developer Consultant having 6.5 years of experience in IT industry including 4+ Years of experience as Hadoop / Spark developer using Big Data technologies like Hadoop ecosystem, Spark ecosystem building scalable, distributed data solutions and driving business required solution. Professional Experience Extensive knowledge in Business Intelligence Data Warehousing Concepts with emphasis on ETL and Life Cycle Development. Experience in Data Warehouse applications testing. A team player with a will-do attitude, phenomenal communication skills & strong user focus.
1. In depth understanding/knowledge of hadoop architecture and various components such as HDFS,Application Master,Node Manager,Resource Manager,Name Node, Data Node & Map Reduce Concepts.
2. Involved in Business requirerment gathering, Technical design documents,business use cases & data mapping.
3. Developed DataPipeline for many Banking Products using sqoop, MR & Hive to extract the data from multiple source system & store the results for downstream consumption in Teradata.
4. Involved in creating Hive tables and loading data & writing hive queries & scripts.
5. Developed Map/Reduce Jobs using PySpark for DataTransformation.
6. Experience working with data formats including XML,CSV,Parquet, etc.
7. Used Jira tracking tool to manage and track the issues reported by QA, prioritize and take action based on the severity.
8. Worked on root cause analysis for all the issues that occur in batch & provide permanent fixes for the issues.
9. Solid knowledge of RDBMS optimizer and performance tuning techniques.
1. Extensive Experience on Teradata database, analyzing business
needs of clients, developing effective and efficient solutions and
ensuring client deliverables with in committed timelines.
2. Experience in Data Warehousing with knowledge on Extraction, Transformation and Loading (ETL) process using Teradata (SQL
Assistant 16.20, Teradata Studio) & IBM WebSphere DataStage
9.1 & 11.5 (Designer, Director and Administrator).
3. Created Teradata objects like Databases, Users, Profiles, Roles, Tables, and Views & Macros.
4. Involved in Teradata Development based on the business
requirement and solution using multiple Teradata Utilities like
Bteq, FLOAD, FEXPORT, MLOAD. Used to run the ETL jobs through Autosys to load data from source to development database.
5. Involved in complete software development life-cycle(SDLC)
including requirements gathering, analysis, design, development,
testing, implementation and deployment.
6. Involved in the analysis of Issues and proposing solutions to
client Involved in analysis of test results and documenting unit
test results.
7. Extensive experience with development, testing, debugging,
implementation, documentation and production support.
8. Extensive experience of CI/CD Tool - Teamcity for package
creation and deployment.
HDFS
Certified in AWS Cloud Practitioner (CLF-C01)
Data Science
Artificial Inteligence
Chess
Marathon
Singing
Certified in AWS Cloud Practitioner (CLF-C01)
Star Performer Awardee for Multiple Product Deliverable's
Mfdm™ AI Aware Certified - TCS Internal
Winner of Texas Instruments India Analog Maker Competition