Skip to content

Commit cbf814a

Browse files
authored
Merge pull request #53834 from MicrosoftDocs/NEW-fabric-modeling
3 new Fabric modeling modules
2 parents d7a24ab + 90851b7 commit cbf814a

69 files changed

Lines changed: 3191 additions & 0 deletions

File tree

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.introduction
3+
title: Introduction
4+
metadata:
5+
title: Introduction
6+
description: "Introduction to choosing analytical data stores in Microsoft Fabric, including lakehouse, warehouse, and eventhouse options."
7+
ms.date: 02/23/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 2
12+
content: |
13+
[!include[](includes/1-introduction.md)]
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.describe-options
3+
title: Describe analytical data store options
4+
metadata:
5+
title: Describe Analytical Data Store Options
6+
description: "Compare the three primary analytical data stores in Microsoft Fabric: lakehouse, warehouse, and eventhouse. Understand decision factors including data format, query language, and write patterns."
7+
ms.date: 02/23/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 6
12+
content: |
13+
[!include[](includes/2-describe-options.md)]
Lines changed: 32 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,32 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.evaluate-lakehouse
3+
title: Evaluate lakehouse capabilities
4+
metadata:
5+
title: Evaluate Lakehouse Capabilities
6+
description: "Evaluate when to choose a lakehouse in Microsoft Fabric. Understand lakehouse architecture, dual Spark and SQL access, schema flexibility, and ideal use cases for data engineering and data science."
7+
ms.date: 02/24/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 6
12+
content: |
13+
[!include[](includes/3-evaluate-lakehouse.md)]
14+
quiz:
15+
title: "Check your understanding"
16+
questions:
17+
- content: "A data science team needs to process 50 TB of mixed structured CSV files and semi-structured JSON web logs using Python notebooks for exploratory analysis and ML model training. Is lakehouse the right choice?"
18+
choices:
19+
- content: "Yes, lakehouse is ideal for this scenario"
20+
isCorrect: true
21+
explanation: "Correct. Mixed data formats (CSV and JSON), Python/Spark notebooks for processing, exploratory analysis with evolving schema, and big data scale (50 TB) are all core lakehouse strengths."
22+
- content: "No, choose a different store"
23+
isCorrect: false
24+
explanation: "Incorrect. Lakehouse is purpose-built for exactly this type of data engineering and data science workload."
25+
- content: "A finance team needs to build a star schema where dimension tables require daily UPDATE operations via T-SQL to maintain slowly changing dimensions. Is lakehouse the right choice?"
26+
choices:
27+
- content: "Yes, lakehouse is ideal for this scenario"
28+
isCorrect: false
29+
explanation: "Incorrect. While lakehouse has a SQL analytics endpoint, it's read-only. You cannot execute UPDATE, DELETE, or MERGE statements via T-SQL against the lakehouse."
30+
- content: "No, choose a different store"
31+
isCorrect: true
32+
explanation: "Correct. The warehouse provides full T-SQL DML support including UPDATE, DELETE, and MERGE operations for maintaining dimensional models with transactional consistency."
Lines changed: 32 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,32 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.evaluate-warehouse
3+
title: Evaluate warehouse capabilities
4+
metadata:
5+
title: Evaluate Warehouse Capabilities
6+
description: "Evaluate when to choose a warehouse in Microsoft Fabric. Understand full T-SQL support, multi-table ACID transactions, schema-on-write governance, and ideal use cases for BI reporting and dimensional modeling."
7+
ms.date: 02/24/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 6
12+
content: |
13+
[!include[](includes/4-evaluate-warehouse.md)]
14+
quiz:
15+
title: "Check your understanding"
16+
questions:
17+
- content: "A BI team with 10+ years of SQL Server experience needs to build dimensional models with daily dimension updates. They need to support complex multi-table joins for Power BI reports. Is warehouse the right choice?"
18+
choices:
19+
- content: "Yes, warehouse is ideal for this scenario"
20+
isCorrect: true
21+
explanation: "Correct. SQL Server expertise translates directly to warehouse T-SQL, dimensional models are optimized in warehouse, transactional updates work for dimension maintenance, and Power BI Direct Lake provides optimal performance."
22+
- content: "No, choose a different store"
23+
isCorrect: false
24+
explanation: "Incorrect. Warehouse is purpose-built for structured BI workloads with T-SQL and dimensional modeling."
25+
- content: "An analytics team works entirely in T-SQL and needs to query data across multiple sources: two different warehouses and Delta tables from a lakehouse. They don't want to copy or move data. Is warehouse the right choice for their queries?"
26+
choices:
27+
- content: "Yes, warehouse is ideal for this scenario"
28+
isCorrect: true
29+
explanation: "Correct. Warehouse cross-database queries support three-part naming to join data across multiple warehouses and lakehouse SQL analytics endpoints without copying data. This is exactly what the team needs."
30+
- content: "No, choose a different store"
31+
isCorrect: false
32+
explanation: "Incorrect. Since the team uses T-SQL and needs to query across warehouses and lakehouses, warehouse's cross-database query capability is ideal."
Lines changed: 32 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,32 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.evaluate-eventhouse
3+
title: Evaluate eventhouse capabilities
4+
metadata:
5+
title: Evaluate Eventhouse Capabilities
6+
description: "Evaluate when to choose an eventhouse in Microsoft Fabric. Understand streaming ingestion, KQL analytics, time-series optimization, and ideal use cases for IoT telemetry and real-time monitoring."
7+
ms.date: 02/24/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 5
12+
content: |
13+
[!include[](includes/5-evaluate-eventhouse.md)]
14+
quiz:
15+
title: "Check your understanding"
16+
questions:
17+
- content: "A reporting team needs to analyze sales data that's updated once per day in batch. They build star schema dimensional models and write complex T-SQL queries with multi-table joins. Is eventhouse the right choice?"
18+
choices:
19+
- content: "Yes, eventhouse is ideal for this scenario"
20+
isCorrect: false
21+
explanation: "Incorrect. Eventhouse is optimized for streaming time-series data with KQL, not batch dimensional modeling with T-SQL joins."
22+
- content: "No, choose a different store"
23+
isCorrect: true
24+
explanation: "Correct. Warehouse is the right choice for batch-updated dimensional models with T-SQL analytics and BI reporting."
25+
- content: "A team collects application log data continuously but only analyzes it weekly to identify trends and errors. They're comfortable learning KQL if it provides benefits. Is eventhouse the right choice?"
26+
choices:
27+
- content: "Yes, eventhouse is ideal for this scenario"
28+
isCorrect: true
29+
explanation: "Correct. Even though analysis is weekly, continuous log ingestion and log data's time-series nature make eventhouse a strong fit. KQL's text parsing and pattern matching work well for log analysis."
30+
- content: "No, choose a different store"
31+
isCorrect: false
32+
explanation: "Incorrect. Continuous log ingestion with time-series characteristics aligns with eventhouse strengths, even if analysis happens weekly rather than real-time."
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.exercise
3+
title: "Case study - Choose data stores for an integrated analytics solution"
4+
metadata:
5+
title: "Case Study - Choose Data Stores for an Integrated Analytics Solution"
6+
description: "Explore how lakehouse, warehouse, eventhouse, and other data stores work together in a complete Fabric architecture through a real-world retail scenario."
7+
ms.date: 02/24/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 8
12+
content: |
13+
[!include[](includes/6-exercise.md)]
Lines changed: 69 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,69 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.knowledge-check
3+
title: Module assessment
4+
metadata:
5+
title: Module Assessment
6+
description: "Test your understanding of choosing data stores in Microsoft Fabric, including decision criteria for lakehouse, warehouse, and eventhouse."
7+
ms.date: 02/23/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
module_assessment: true
12+
durationInMinutes: 3
13+
quiz:
14+
questions:
15+
- content: "Which analytical data store in Microsoft Fabric provides full multi-table ACID transaction support through T-SQL?"
16+
choices:
17+
- content: "Lakehouse"
18+
isCorrect: false
19+
explanation: "Incorrect. The lakehouse supports single-table ACID transactions through Delta Lake, but it doesn't support multi-table transactions through T-SQL. The lakehouse's SQL analytics endpoint is read-only."
20+
- content: "Warehouse"
21+
isCorrect: true
22+
explanation: "Correct. The warehouse provides full multi-table ACID transaction support through T-SQL, including INSERT, UPDATE, DELETE, and MERGE operations across multiple tables."
23+
- content: "Eventhouse"
24+
isCorrect: false
25+
explanation: "Incorrect. The eventhouse is optimized for append-only streaming ingestion and doesn't support transactional UPDATE or DELETE operations."
26+
- content: "A data science team needs to explore a mix of structured transaction data and semi-structured web logs using Python notebooks. Which data store is the best fit?"
27+
choices:
28+
- content: "Lakehouse"
29+
isCorrect: true
30+
explanation: "Correct. The lakehouse supports both structured and semi-structured data, and provides native Apache Spark access through Python notebooks for data exploration and machine learning."
31+
- content: "Warehouse"
32+
isCorrect: false
33+
explanation: "Incorrect. The warehouse is designed for structured data and T-SQL development. It doesn't provide Spark notebook access for Python-based data exploration."
34+
- content: "Eventhouse"
35+
isCorrect: false
36+
explanation: "Incorrect. The eventhouse is optimized for time-series and streaming data with KQL as the primary query language. It isn't designed for mixed-format data exploration with Python notebooks."
37+
- content: "What is the primary query language used by the eventhouse for time-series analytics?"
38+
choices:
39+
- content: "T-SQL"
40+
isCorrect: false
41+
explanation: "Incorrect. While the eventhouse supports a subset of T-SQL, KQL (Kusto Query Language) is the primary query language designed specifically for time-series analysis with built-in operators for aggregations, anomaly detection, and pattern matching."
42+
- content: "Spark SQL"
43+
isCorrect: false
44+
explanation: "Incorrect. Spark SQL is the query language used in lakehouse notebooks. The eventhouse uses KQL as its primary query language."
45+
- content: "KQL (Kusto Query Language)"
46+
isCorrect: true
47+
explanation: "Correct. KQL is the primary query language for the eventhouse, purpose-built for time-series analysis with operators for time-window aggregations, anomaly detection, and pattern matching."
48+
- content: "An organization needs to build a star schema with dimension tables that require frequent updates. Which data store best supports this requirement?"
49+
choices:
50+
- content: "Warehouse"
51+
isCorrect: true
52+
explanation: "Correct. The warehouse provides full T-SQL DML support including UPDATE, DELETE, and MERGE operations with multi-table ACID transactions, which is essential for maintaining slowly changing dimensions in a star schema."
53+
- content: "Lakehouse"
54+
isCorrect: false
55+
explanation: "Incorrect. While the lakehouse supports single-table transactions through Delta Lake, it doesn't provide the full T-SQL DML surface (UPDATE, DELETE, MERGE) through T-SQL. The SQL analytics endpoint is read-only."
56+
- content: "Eventhouse"
57+
isCorrect: false
58+
explanation: "Incorrect. The eventhouse is designed for append-optimized time-series data, not for relational dimensional models that require frequent row-level updates."
59+
- content: "Which feature allows data in one Fabric data store to be accessed from another store without copying or moving the data?"
60+
choices:
61+
- content: "Cross-database queries and shortcuts"
62+
isCorrect: true
63+
explanation: "Correct. Shortcuts let you reference data in one store from another without duplication, and cross-database queries in the warehouse let you join data from multiple warehouses and lakehouse SQL analytics endpoints using three-part naming."
64+
- content: "Data pipelines"
65+
isCorrect: false
66+
explanation: "Incorrect. Data pipelines move or copy data between sources and destinations. Shortcuts and cross-database queries provide access to data in place without copying it."
67+
- content: "Streaming ingestion"
68+
isCorrect: false
69+
explanation: "Incorrect. Streaming ingestion is a method for loading real-time data into an eventhouse. It doesn't provide cross-store data access without copying."
Lines changed: 13 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,13 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.choose-data-store-fabric.summary
3+
title: Summary
4+
metadata:
5+
title: Summary
6+
description: "Review key concepts for choosing analytical data stores in Microsoft Fabric, including when to use lakehouse, warehouse, and eventhouse for different business scenarios."
7+
ms.date: 02/23/2026
8+
author: angierudduck
9+
ms.author: anrudduc
10+
ms.topic: unit
11+
durationInMinutes: 2
12+
content: |
13+
[!include[](includes/8-summary.md)]
Lines changed: 16 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,16 @@
1+
Microsoft Fabric provides multiple analytical data stores, each designed for different data patterns and workloads. Choosing the right store for your scenario is one of the most important decisions you make when building an analytics solution. The right choice affects how you ingest data, how your team queries it, and how well your solution adapts as requirements evolve.
2+
3+
The three primary analytical data stores in Fabric are the lakehouse, the warehouse, and the eventhouse. Each one stores data in OneLake using an open format, but they differ in query language support, write capabilities, and the types of workloads they handle best.
4+
5+
Suppose you're a data professional at a retail organization that's adopting Microsoft Fabric. Your team needs to build analytics solutions for several business groups. The sales team wants structured reporting with complex joins. The data science team needs to explore a mix of transaction data and web clickstream logs. The operations team monitors IoT sensor data from distribution centers in real time. Each group has different data types, query preferences, and performance requirements.
6+
7+
Your task is to evaluate Fabric's analytical data stores and recommend the right one for each business group. To do that, you need to understand what each store does well, when to use it, and when a different option is a better fit.
8+
9+
In this module, you:
10+
11+
- Survey the three analytical data store options in Microsoft Fabric: lakehouse, warehouse, and eventhouse.
12+
- Evaluate the strengths and ideal use cases for each data store.
13+
- Apply a decision framework to match data characteristics and team skills to the right store.
14+
- Practice choosing the appropriate data store for real-world business scenarios.
15+
16+
By the end of this module, you're able to evaluate lakehouse, warehouse, and eventhouse capabilities and confidently choose the appropriate data store for a given business scenario.

0 commit comments

Comments
 (0)