OnBenchMark Logo


Software Engineer
Location India
Total Views133
back in time
Member since30+ Days ago
 back in time
Contact Details
phone call {{contact.cdata.phone}}
phone call {{contact.cdata.email}}
Candidate Information
  • User Experience
    Experience 8 Year
  • Cost
    Hourly Rate$8
  • availability
  • work from
    Work FromOffsite
  • check list
    CategoryInformation Technology & Services
  • back in time
    Last Active OnJune 22, 2024
Key Skills
MongoDBSQLInformaticaSparkScalaHiveSqoopETLUnixShell Scripting
  • Having 8 years of total experience in Oracle,Informatica PowerCenter and Bigdata Technologies.
  • 3+ Years of experience in Big Data-Hadoop Frameworks (Spark,Scala,Hive,Sqoop,HDFS).
  • Strong Knowledge of Hadoop Architecture and Daemons like NameNode,DataNode,Resource Manager, Node Manager etc.
  • Implemented pre-defined operators in Spark such as map,flatmap,filter,reduceByKey,groupByKey.
  • Used Scala sbt to develop scala coded spark projects and executed using spark-submit.
  • Hands on experience in writing spark programs using scala to handle different data sources using SparkSQL.
  • Knowledge in Scala concepts like patteren matching, Anonymous function ,Abstract class, Traits etc.
  • Experience in developing ad-hoc reports using Hive
  • Extensive knowledge of creating managed tables and external tables in Hive Eco system including ORC and Parquet tables.
  • Responsible for user management module and domain monitoring and management using Sqoop scripts in order to make the interaction between Hive and Oracle Database.
  • Worked on automation and Job Monitoring using the scheduling tool Autosys.
  • Used ETL process to Extract, Transform and Load the data into stage area and data warehouse

Using Informatica Power Center.

  • Extensively used ETL to load data from Flat files and databases like Oracle.
  • Experience in creating different kinds of Partitions on tables and indexes for managing large tables..
  • Extensively worked on UNIX and Shell Scripting
  • Performing data cleansing, scrubbing and validating the data before loading the data into Data lakes.
  • Adequate knowledge and working experience with Agile and waterfall methodologies.
  • Debug, deploy and maintain Sqoop scripts to extract data to/from RDBMS to Hadoop.
  • Strong knowledge in Hive Windowing Functions and Oracle Analytic Functions.
  • Strong debugging and critical thinking ability with good understanding of frameworks advancement in methodologies and strategies.

Copyright© Cosette Network Private Limited All Rights Reserved
Submit Query
WhatsApp Icon