· Overall 3.4 years of experience with Big data technologies like Hadoop ecosystem.
Hands on experience in Data Ingestion, Transformation, Analytics using PySpark framework, Hadoop eco systems components.
Extensively used SQL, Numpy, Pandas, PySpark, Python for Data Analysis and Model building.
Hands on experience in Hadoop components like PySpark, HDFS, Hive,PigSqoop, and in-depth knowledge on Hadoop architecture.
Good experience in working with Hive data warehouse tool-creating tables, data distribution by implementing partitioning and bucketing, writing and optimizing the HiveQL queries.
Experience on writing Queries, Stored procedures and functions by using SQL.
Good knowledge in DBMS like MS SQL Server.
Extensively worked on Sqoop, Hadoop, Hive and PySpark, ETL and Data Processing systems.
Having Good Understanding of Dataware house.
Worked in all phases of software development life cycle.