Scala code to convert CSV files stored in Azure Blob Storage to Parquet and store into Azure Storage, using Data bricks notebook and ARM template to run the notebook as a Azure Data Factory Job
-
Updated
Feb 26, 2023 - Scala
Scala code to convert CSV files stored in Azure Blob Storage to Parquet and store into Azure Storage, using Data bricks notebook and ARM template to run the notebook as a Azure Data Factory Job
The future of sustainability and training: involvement and performance of companies and strategic suppliers
Data pipeline project (ELT using Microsoft Azure)
This is a project where data ingestion, data transformation, data preparation and other data activities including Azure SQL. Making pipelines production ready, monitoring and CI/CD implementation.
Bugs In Cloud - Elenco dei Video
Contains solutions/versions of Batch Data Pipelines created on Azure Data Factory
Tokyo Olympics Data Analysis: Creating a ETL pipeline using Azure Data Factory to ingest data, transform it using Azure Databricks and querying and building reports using tools like Synapse Analytics and PowerBI
A Covid-19 Project on Azure Cloud
The aim of this project is to build a cost efficient Data Warehouse on Amazon's Retail sales data and perform Customer lifetime value analyses.
This project presents a sophisticated data-driven web application that integrates React for frontend visualization, NodeJS for backend data retrieval, and Microsoft's Cosmos DB for data storage. Leveraging the fault tolerance, partitioning, replication, and global distribution advantages of Cosmos DB.
Repository created for programming and development in the Azure Data Enginner.
Microsoft Internship Program: During this Internship, I have worked on projects related to some of the machine learning algorithms. And deployed the model using Microsoft Azure.
Integration of Covid-19 data utilising Azure Data Factory to perform data ingestion, transformation and storage activities. The goal of this guided project was to become familiar with Microsoft Azure technologies, including; Azure Data Factory(ADF), Azure Data Lake Storage Gen2, Azure SQL Database, Azure Blob Storage, Dataflow, Databricks, etc.
The aim of this project is to build a cost efficient Data Warehouse on Amazon's Retail sales data and perform Customer lifetime value analyses
Azure for End to End Data Science Project
Implemented an end-to-end Azure data engineering solution to process Tokyo Olympics 2021 data, encompassing extraction, transformation, analytics, and visualization.
Copying data from Amazon S3 bucket to Azure Blob container by using Azure Data Factory pipeline. This Data is mounted to Databricks and further analysis is done using Spark SQL.
Add a description, image, and links to the azuredatafactory topic page so that developers can more easily learn about it.
To associate your repository with the azuredatafactory topic, visit your repo's landing page and select "manage topics."