OnBenchMark Logo

Sudheer (RID : 15cyvlmzzq374)

designation   AWS Data Engineer

location   Location : Delhi, India,

experience   Experience : 8 Year

rate   Rate: $18 / Hourly

Availability   Availability : Immediate

Work From   Work From : Offsite

designation   Category : Information Technology & Services

Shortlisted : 1
Total Views : 56
Key Skills
AWS SQL Python Performance Tuning Database Modeling OracleDB PostgresDB Airflow Snowflake AWS & GCP Dynamic & Advanced SQL DWH Concepts Data Pipeline Design
Discription

EXPERIENCE

NOV2021-PRESENT

  • Ledacriticaldataengineeringproject,receivingrawdatainJSONformatandimplementing a streamlined workflow. Developed AWS Lambda functions to clean andtransformthefilesintoParquet,enhancingdatastorageandqueryperformance.
  • Orchestrated data loading into Redshift using AWS Glue, ensuring seamless integrationandenablingefficientanalysisandreporting.Theprojectinvolveddesigninganautomatedprocesstohandlelargevolumesofdata,optimizingingestionspeed.
  • Spearheaded the development of a robust data model based on Dimensional Modelingprinciples.CreatedmultipledimensionsandfactswithinRedshift,enablingcomprehensive and insightful data analysis for business stakeholders across variousdomains.
  • Identifiedprocessimprovementopportunitiesandimplementeddatavalidationmechanisms during the cleansing phase, ensuring data accuracy and integrity throughouttheprojectlife-cycle.Thisenhancedthereliabilityofdownstreamanalyticsanddecision-makingprocesses.
  • Optimized the ETL pipeline by leveraging parallel processing techniques and resourceallocation strategies, significantly reducing data transformation and loading times. Thisimprovementenhancedoverallsystem performance and increased data processingefficiency.
  • Implemented incremental loading techniques to process and load only new or modifieddata,minimizingprocessingtimeandincreasingtheproject'sscalability.Thisapproach

 

Reducedunnecessaryprocessing,enablingfasterdataavailabilityforanalyticsandreporting.

 

XYZ,Bangalore—DataEngineer

SEP2018-NOV2021

  • Ledreal-timedataingestionusingGooglePub/SubandSnow-pipe,automatingdataloadingintoSnowflake.
  • OrchestratedPub/SubandSnowflakeintegrationforoptimizedandscalabledataingestion.
  • DesignedacomprehensivedimensionalmodelinSnowflakeforadvancedanalytics.
  • Implemented proactive monitoring for efficient pipeline management and enhanced datareliability.
  • Collaboratedtodelivertailoreddatamodelsalignedwithbusinessneeds.
  • Ensureddataaccuracythroughrigorousvalidationpractices.
  • Establishedcleardocumentationanddatalineageforeffectivedatagovernance.
Project Details
Title : AWS Project
Duration : 24 (Month)
role and responsibileties :
  • Led a critical data engineering project, receiving raw data in JSON format and implementing a streamlined workflow. Developed AWS Lambda functions to clean and transform the files into Parquet, enhancing data storage and query performance.
  • Orchestrated data loading into Redshift using AWS Glue, ensuring seamless integration and enabling efficient analysis and reporting. The project involved designing an automated process to handle large volumes of data, optimizing ingestion speed.
  • Spearheaded the development of a robust data model based on Dimensional Modeling principles. Created multiple dimensions and facts within Redshift, enabling comprehensive and insightful data analysis for business stakeholders across various domains.
  • Identified process improvement opportunities and implemented data validation mechanisms during the cleansing phase, ensuring data accuracy and integrity throughout the project life-cycle. This enhanced the reliability of downstream analytics and decision-making processes.
  • Optimized the ETL pipeline by leveraging parallel processing techniques and resource allocation strategies, significantly reducing data transformation and loading times. This improvement enhanced overall system performance and increased data processing efficiency.
  • Implemented incremental loading techniques to process and load only new or modified data, minimizing processing time and increasing the project's scalability. 
Description :

 Proficient in database modeling, data warehousing, and Unix. Skilled in cloud platforms (AWS, GCP) and proficient with Snowflake for efficient data processing. Diverse industry experience in Banking, Media, and Investment Banking. Strong ability to design, develop, and optimize data pipelines and architectures. Collaborative team player delivering high-quality projects on time and within budget. A results-oriented professional driving efficiency and enabling data-driven decision-making.


 
Matching Resources
My Project History & Feedbacks
Copyright© Cosette Network Private Limited All Rights Reserved
Submit Query
WhatsApp Icon
Loading…

stuff goes in here!