You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: learn-pr/wwl-databricks/select-and-configure-compute/2-choose-appropriate-compute-type.yml
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ title: Choose an appropriate compute type
4
4
metadata:
5
5
title: Choose an Appropriate Compute Type
6
6
description: Learn how to choose the right compute type for your Azure Databricks workloads, comparing serverless, classic, SQL warehouses, instance pools, and job compute options.
description: Learn how to configure Azure Databricks compute performance settings including node types, autoscaling, termination, and instance pools to balance cost and performance.
Copy file name to clipboardExpand all lines: learn-pr/wwl-databricks/select-and-configure-compute/4-configure-compute-feature-settings.yml
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ title: Configure compute features
4
4
metadata:
5
5
title: Configure Compute Features
6
6
description: Learn how to configure Azure Databricks compute features including Photon acceleration, Databricks runtime versions, and machine learning environments for optimal workload performance.
Copy file name to clipboardExpand all lines: learn-pr/wwl-databricks/select-and-configure-compute/5-install-libraries-for-compute.yml
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ title: Install libraries for compute
4
4
metadata:
5
5
title: Install Libraries for Compute
6
6
description: Learn how to install libraries on Azure Databricks compute resources using package repositories, workspace files, Unity Catalog volumes, and init scripts.
Copy file name to clipboardExpand all lines: learn-pr/wwl-databricks/select-and-configure-compute/6-configure-compute-access-perms.yml
+1-1Lines changed: 1 addition & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -4,7 +4,7 @@ title: Configure compute access
4
4
metadata:
5
5
title: Configure Compute Access
6
6
description: Learn how to configure access permissions for Azure Databricks compute resources, including permission levels, access modes, dedicated group access, and workspace-level entitlements.
Every Azure Databricks workload runs on compute resources, but choosing the wrong compute type or configuration leads to unnecessary costs, poor performance, or blocked functionality. **Serverless compute** starts in seconds but doesn't support RDD APIs. **Classic compute** offers complete flexibility but requires more management overhead. **SQL warehouses** excel at analytical queries while **job clusters** optimize for automated workflows. Understanding these differences helps you match compute to workload requirements.
2
4
3
5
Beyond selecting a compute type, configuration decisions shape how your workload performs. **Node types** determine processing capacity and memory availability. **Autoscaling** balances cost and responsiveness. **Access permissions** control who can use compute resources while **library installations** provide the dependencies your code needs. Each configuration choice affects multiple dimensions—performance, cost, security, and operational complexity.
**Serverless compute** is managed entirely by Azure Databricks. You don't provision or configure infrastructure—Azure Databricks automatically allocates and scales resources based on your workload demands. These resources run in **Databricks' Azure subscription, not yours**, which means no virtual machines or networking components appear in your subscription.
8
10
9
11
With serverless compute, startup typically takes 2-6 seconds. The platform scales up rapidly when query volume increases and scales down during idle periods to minimize costs. This eliminates the need to estimate capacity or manage cluster configurations.
@@ -21,6 +23,8 @@ However, serverless has limitations. You can't use **RDD APIs** (Resilient Distr
**Classic compute** gives you full control over cluster configuration. You create, size, and manage compute resources that run directly in **your Azure subscription**, giving you visibility and control over the underlying infrastructure.
25
29
26
30
Classic compute supports two access modes that determine how users interact with the cluster:
@@ -43,6 +47,8 @@ This compute type fits workloads that need features unavailable in serverless, r
**SQL warehouses** are compute resources optimized specifically for SQL queries, analytics, and business intelligence. They come in three types, each with different performance characteristics.
47
53
48
54
**Serverless SQL warehouses** offer optimal performance and cost efficiency. They start in 2-6 seconds, use Intelligent Workload Management to predict query resource needs, and scale clusters dynamically based on demand. Photon and Predictive IO accelerate query execution. Choose serverless SQL warehouses for most SQL workloads—BI dashboards, ETL jobs, and ad hoc analysis.
@@ -55,6 +61,8 @@ All SQL warehouse types optimize for SQL execution patterns, but serverless offe
**Instance pools** maintain a set of idle virtual machine instances ready for immediate use. When you create a cluster from a pool, startup time decreases because Databricks allocates instances from the pool instead of requesting new ones from Azure.
59
67
60
68
Pools reduce startup time from minutes to under a minute in many cases. You configure the minimum number of idle instances to keep warm and the maximum pool capacity. When clusters release instances, those instances return to the pool for reuse.
@@ -67,6 +75,8 @@ Configure pools with spot instances for worker nodes to reduce costs, but use on
**Job compute** refers to clusters optimized for automated workflows rather than interactive development. You configure job compute through cluster policies that enforce best practices for production workloads.
71
81
72
82
Job clusters terminate automatically after completing their tasks, preventing unnecessary costs from idle resources. When you configure a job, you choose between serverless and classic job compute.
@@ -80,6 +90,8 @@ The Job Compute policy in Azure Databricks offers a template for creating produc
Different compute types suit different scenarios. The following table compares key characteristics to help you make informed decisions:
84
96
85
97
| Compute type | Recommended for | Startup time | Management overhead | Cost efficiency | Key limitation |
@@ -98,7 +110,7 @@ Different compute types suit different scenarios. The following table compares k
98
110
99
111
Start your decision-making process by identifying your workload characteristics. The following diagram illustrates a decision flow to help you select the appropriate compute type:
100
112
101
-

113
+

0 commit comments