Total 7 years of experience in Big data and ETL / Business Intelligence technology
2 Years of experience on Big data Technologies (Hadoop, Spark, Scala, Python, Hive, Airflow, GCP) Worked on Google Cloud Platform for running Spark based Dataproc cluster jobs.
Worked on TB size of data Processing using Spark distributing processing.
Worked on Elastic Search for uploading TB size of data.
Embedded Fast-text python code in Spark Scala using JEP interpreter.
Worked on Akka based customer allocation of promotions.
Creating Hive Query for business given queries.
Experience of Data scraping with Python.
Trained Junior Team members in use of Big Data Technologies as well as informatica PowerCenter. Also Conducted Functional and Technical knowledge sharing sessions for the team.
Expert in ETL technologies (Informatica PowerCenter, PL/SQL, Unix, Control-M)
Experience in Integrating Sales Force data To EDW using Informatica webservices
Expert in SQL and writing complex queries on Oracle 12c, MS SQL.
Knowledge in Data Warehousing Dimensional Modeling, Data Warehousing concepts.
Responsible for Extraction, Transformation, and Loading of data from multiple sources into Data Warehouse.
Designed and developed Mappings, from varied Transformations logic like Unconnected and Connected Lookups, Router, Filter, Expression, Aggregator, Joiner, Update strategy and more.
Experience of using Web Services functionality of informatica to trigger jobs on multiple nodes in parallel.