Kalaneni

  • Data Engineer
  • Hyderabad

Rate

$ 5.00 (Hourly)

Experience

2 Years

Availability

Immediate

Work From

Offsite

Skills

Data EngineerAzure DatabricksPythonSQLPySpark

Description

WORK EXPERIENCE Software Engineer 1, Microsoft Azure MAQ Software, Hyderabad Duration: Aug 22 – Present Technologies/Tools: Azure Data Factory, Azure Databricks, SQL, Azure Data Lake Storage, Azure Synapse, PySpark Description: An orchestrating framework to bring the data from external data sources like SFTP, Google Storage, Amazon S3 bucket into ADLS till Synapse and make it available to downstream end users.

• Preparing Requirements Analysis and Data Architecture Design.

• Building the pipelines to copy the data from source to destination in ADF.

• Automated the process of Linked Service and Dataset creation on the source as well as destination servers.

• Transforming data in Azure Data Factory with ADF Transformations.

• Scheduling pipelines and monitoring the data movement from source to destination.

• Creating multiple types of triggers to schedule the pipelines.

• Working on multiple data sources PRD’s .

• Working with Databricks Notebooks to transform data in alignment with specific use cases using Pyspark and SQL.

• Creating, Configuring and Monitoring Interactive clusters and job clusters.

• Automating Mail Notification Alert using Databricks Notebook through API calls.

• Creating STTMs (Source to Target Mapping) for Data Modelling.

• Build and Run Continuous Integration and Continuous Delivery (CI/CD) Pipeline in Azure Data Factory.

• Building Databricks Validation Notebooks to monitor the fetched and transformed Data.

• Creating Data assets and Data Quality checks in Internal tool.

• Experience with agile Delivery Methodologies. Software Engineer, Microsoft Azure Capgemini, Hyderabad Duration: Aug 21 – Aug 22 Technologies/Tools: Azure Data Factory, Azure SQL, Azure Databricks Description: Worked on ETL pipelines to automate the task of fetching the data from multiple sources and loading into multiple destinations.

• Worked with Python and SQL to write the logic to transform the data.

• Building the pipelines to copy the data from source to destination in ADF.

• Creating Linked Services and Multiple datasets on the source as well as destination servers.

• Scheduling pipelines and monitoring the data movement from source to destination.

Submit Query icon