As businesses continue to generate massive amounts of data, the need for efficient data management solutions becomes increasingly important. This is where a data lake house comes in - an hybrid solution that combines the best features of a datalake and a data warehouse.
Part 1 - Building a Data Lakehouse using Azure Data explorer
We explored how to build a data lakehouse using Azure Data Explorer (ADX) where the data flows from Azure SQL DB using Change Data Capture (CDC) through Azure Data Factory and events flowing from events hub.
This article is Part 2 in the series, here we will deploy this solution using Bicep, a powerful infrastructure as code (IaC) tool from Microsoft. With this guide, you'll be able to create a data lakehouse that can handle large volumes of data and provide valuable insights for your business.
https://github.com/denisa-ms/azure-data-and-ai-examples/tree/master/adx-datalakehouse
NOTE: This takes time so be patient
The code here creates the following entities
Contains an Azure SQL database with the Adventure works sample data.
Contains 2 data pipelines:
Contains a hub called “clicks-stream” that streams click events into ADX table bronzeClicks
In order to run this demo, you should:
Select file
Click start Ingestion.
We are done!
We have products and orders from our operational DB (Azure SQL) and events coming from a stream in events hub.
In this demo I chose to add synthetic events using one-click ingestion, but you can create events and publish them to Events hub and they will be ingested using streaming ingestion to the bronzeClicks table.
I hope you enjoyed this
Thanks
Denise
You must be a registered user to add a comment. If you've already registered, sign in. Otherwise, register and sign in.