There are 2 repositories under datafactory topic.
DataOps for the Modern Data Warehouse on Microsoft Azure. https://aka.ms/mdw-dataops.
Terraform script to deploy almost all Azure Data Services
Threat Detection and Visualization
Azure Container Instances Proxy implemented in Azure Function App (Consumption Plan)
Generic Pipelines / Templates for Data Factory / Synapse Pipelines w.r.t Different MSFT Offering Integrations / Use Cases
A data pipeline project build on databricks and azure to demostrate lifecycle of a cloud data project.
This pipeline is an ETL application, which aims to collect data for a few minutes from Binance's open API for cryptocurrencies: BTCUSDT, ADAUSDT, ETHUSDT, BNBUSDT and LTCUSDT.
Proceso ETL
The ADF Universal Framework is an open-source project designed to provide a comprehensive and flexible solution for building scalable and efficient data integration workflows using Azure Data Factory (ADF).
Explore the Tokyo Olympics data journey! We ingested a GitHub CSV into Azure via Data Factory, stored it in Data Lake Storage Gen2, performed transformations in Databricks, conducted advanced analytics in Azure Synapse, and visualized insights in Synapse or Power BI.
Tokyo Olympic Azure Data Engineering Project
Repository for Azure Data Factory (ADF) Custom Activity to dynamically create and process Azure Analysis Service (AAS) Tabular Model Partitions
This application shows how to use the Azure .NET library to create and execute Azure Data Factory from C# code.
Azure Data Factory template to refresh Power BI Dataset
This Project Extracts supply chain data from csv file having 180k records and more than 40 columns from the Azure Datalake Gen2 storage account and do some dataanalysis with Python(Pandas) to find the top 3 countries and filtered the data for top 3 countries and finally transferred it to 3 files in datalake again by creating ETL pipeline in ADF.
This repository serves as a sample demonstration of implementing CI/CD practices on Azure Data Factory using Azure DevOps. It provides an example approach to configure a pipeline for continuous integration and continuous deployment (CI/CD) workflows on Azure Data Factory
Azure Data Landing Zone - IAC , Config, Code
Interesting_programs_written_in_Python_language
This project showcase the following: 1. REST Connection to CHAT GPT 2. Logging system to Azure file system, storage app and insights. 3. Azure Data Factory Rest implementation using MS Graph client.
End to end data reporting project using Azure services like Azure Data Factory for data orchestration, Azure Synapse Analytics for data warehousing, Databricks for data transformations, and Power BI for intuitive data visualization and reporting.
Efetuar o download de arquivos da web com Python. Inserir dados de um dataframe na cloud Azure com Azure SQL Database. Efetuar transformações nos dados com Azure Data Factory.
This project leverages Azure Cloud services like Azure Data Factory, Azure Databricks, and Synapse Analytics to execute a data engineering workflow. Utilizing data sourced from the Olympic API on GitHub, it involves extracting raw data into Azure Data Lake Storage, transforming it with PySpark on Azure Databricks, and analyzing the transformed data
PsDataFactory is a PowerShell module to generate DataFactory pipelines and datasets.
Data integration with ADF
Data engineering with Azure services
Repository to connect github to Azure Data Factory instance for CI/CD pipeline
Datas Engineer Exploration of the 2021 Tokyo Olympic data using a variety of tools and technologies, with a primary focus on Azure services, Leveraging Azure Data Factory, Data Lake Gen 2, Synapse Analytics, and Azure Databricks to perform data engineering tasks, analysis, and visualization.
O objetivo deste projeto foi aplicar os conhecimentos nas ferramentas de extração e tratamento de dados da plataforma Azure.
DataOps for the Modern Data Warehouse on Microsoft Azure. https://aka.ms/mdw-dataops.
End-to-End data engineering project with Azure Databricks as cloud service and Tokyo olympic data