You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: articles/hdinsight-aks/monitor-with-prometheus-grafana.md
+7-3Lines changed: 7 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -3,7 +3,7 @@ title: Monitoring with Azure Managed Prometheus and Grafana
3
3
description: Learn how to use monitor With Azure Managed Prometheus and Grafana
4
4
ms.service: hdinsight-aks
5
5
ms.topic: how-to
6
-
ms.date: 08/29/2023
6
+
ms.date: 10/27/2023
7
7
---
8
8
9
9
# Monitoring with Azure Managed Prometheus and Grafana
@@ -23,8 +23,8 @@ This article covers the details of enabling the monitoring feature in HDInsight
23
23
* An Azure Managed Prometheus workspace. You can think of this workspace as a unique Azure Monitor logs environment with its own data repository, data sources, and solutions. For the instructions, see [Create a Azure Managed Prometheus workspace](../azure-monitor/essentials/azure-monitor-workspace-manage.md).
24
24
* Azure Managed Grafana workspace. For the instructions, see [Create a Azure Managed Grafana workspace](../managed-grafana/quickstart-managed-grafana-portal.md).
25
25
* An [HDInsight on AKS cluster](./quickstart-create-cluster.md). Currently, you can use Azure Managed Prometheus with the following HDInsight on AKS cluster types:
26
-
* Apache Spark
27
-
* Apache Flink
26
+
* Apache Spark™
27
+
* Apache Flink®
28
28
* Trino
29
29
30
30
For the instructions on how to create an HDInsight on AKS cluster, see [Get started with Azure HDInsight on AKS](./overview.md).
@@ -164,3 +164,7 @@ You can use the Grafana dashboard to view the service and system. Trino cluster
164
164
1. View the metric as per selection.
165
165
166
166
:::image type="content" source="./media/monitor-with-prometheus-grafana/view-output.png" alt-text="Screenshot showing how to view the output." border="true" lightbox="./media/monitor-with-prometheus-grafana/view-output.png":::
167
+
168
+
## Reference
169
+
170
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](./trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
[Azure HDInsight on AKS](../overview.md) is a managed cloud-based service for big data analytics that helps organizations process large amounts data. This tutorial shows how to use Delta Lake scenario in Azure HDInsight on AKS Spark cluster.
13
+
[Azure HDInsight on AKS](../overview.md) is a managed cloud-based service for big data analytics that helps organizations process large amounts data. This tutorial shows how to use Delta Lake in Azure HDInsight on AKS with Apache Spark™ cluster.
14
14
15
15
## Prerequisite
16
16
17
-
1. Create an [Azure HDInsight on AKS Spark cluster](./create-spark-cluster.md)
17
+
1. Create an [Apache Spark™ cluster in Azure HDInsight on AKS](./create-spark-cluster.md)
1. Run Delta Lake scenario in Jupyter Notebook. Create a Jupyter notebook and select "Spark" while creating a notebook, since the following example is in Scala.
22
22
@@ -33,7 +33,7 @@ ms.date: 08/29/2023
33
33
34
34
### Provide require configurations for the delta lake
35
35
36
-
Delta Lake Spark Compatibility matrix - [Delta Lake](https://docs.delta.io/latest/releases.html), change Delta Lake version based on Spark Version.
36
+
Delta Lake with Apache Spark Compatibility matrix - [Delta Lake](https://docs.delta.io/latest/releases.html), change Delta Lake version based on Apache Spark Version.
:::image type="content" source="./media/azure-hdinsight-spark-on-aks-delta-lake/data-after-each-data-load.png" alt-text="Screenshot KPI data after each data load." border="true" lightbox="./media/azure-hdinsight-spark-on-aks-delta-lake/data-after-each-data-load.png":::
218
218
219
+
## Reference
220
+
221
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
Azure HDInsight on AKS is a managed cloud-based service for big data analytics that helps organizations process large amounts data. This tutorial shows how to use configuration management in Azure HDInsight on AKS Spark cluster.
12
+
Azure HDInsight on AKS is a managed cloud-based service for big data analytics that helps organizations process large amounts data. This tutorial shows how to use configuration management in Azure HDInsight on AKS with Apache Spark™ cluster.
13
13
14
-
Configuration management is used to add specific configurations into the spark cluster.
14
+
Configuration management is used to add specific configurations into the Apache Spark cluster.
15
15
16
16
When user updates a configuration in the management portal the corresponding service is restarted in rolling manner.
17
17
@@ -62,5 +62,9 @@ When user updates a configuration in the management portal the corresponding ser
62
62
> Selecting **Save** will restart the clusters.
63
63
> It is advisable not to have any active jobs while making configuration changes, since restarting the cluster may impact the active jobs.
64
64
65
+
## Reference
66
+
67
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
68
+
65
69
## Next steps
66
70
*[Library management in Spark](./library-management.md)
1. Create a new Notebook and select type as **pyspark**.
30
30
1. Copy the workspace and Lakehouse names into your notebook and build your OneLake URL for your Lakehouse. Now you can read any file from this file path.
1. Test that your data was successfully written by checking in your Lakehouse or by reading your newly loaded file.
41
+
42
+
## Reference
43
+
44
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
Apache Spark is a parallel processing framework that supports in-memory processing to boost the performance of big-data analytic applications.
13
+
Apache Spark™ is a parallel processing framework that supports in-memory processing to boost the performance of big-data analytic applications.
14
14
15
-
Spark provides primitives for in-memory cluster computing. A Spark job can load and cache data into memory and query it repeatedly. In-memory computing is faster than disk-based applications, such as Hadoop, which shares data through Hadoop distributed file system (HDFS). Spark allows integration with the Scala and Python programming languages to let you manipulate distributed data sets like local collections. There's no need to structure everything as map and reduce operations.
15
+
Apache Spark™ provides primitives for in-memory cluster computing. A Spark job can load and cache data into memory and query it repeatedly. In-memory computing is faster than disk-based applications, such as Hadoop, which shares data through Hadoop distributed file system (HDFS). Apache Spark allows integration with the Scala and Python programming languages to let you manipulate distributed data sets like local collections. There's no need to structure everything as map and reduce operations.
16
16
17
17
:::image type="content" source="./media/spark-overview/spark-overview.png" alt-text="Diagram showing Spark overview in HDInsight on AKS.":::
18
18
19
19
20
-
## HDInsight Spark in AKS
20
+
## Apache Spark cluster with HDInsight on AKS
21
21
Azure HDInsight is a managed, full-spectrum, open-source analytics service for enterprises.
22
22
23
-
Apache Spark in Azure HDInsight is the managed spark service in Microsoft Azure. With Apache Spark on AKS in Azure HDInsight, you can store and process your data all within Azure. Spark clusters in HDInsight are compatible with or [Azure Data Lake Storage Gen2](../../storage/blobs/data-lake-storage-introduction.md), allows you to apply Spark processing on your existing data stores.
23
+
Apache Spark™ in Azure HDInsight on AKS is the managed spark service in Microsoft Azure. With Apache Spark in Azure HDInsight on AKS, you can store and process your data all within Azure. Spark clusters in HDInsight are compatible with or [Azure Data Lake Storage Gen2](../../storage/blobs/data-lake-storage-introduction.md), allows you to apply Spark processing on your existing data stores.
24
24
25
25
The Apache Spark framework for HDInsight on AKS enables fast data analytics and cluster computing using in-memory processing. Jupyter Notebook lets you interact with your data, combine code with markdown text, and do simple visualizations.
26
26
27
-
Spark on AKS in HDInsight composed of multiple components as pods.
27
+
Apache Spark on AKS in HDInsight composed of multiple components as pods.
28
28
29
29
## Cluster Controllers
30
30
31
31
Cluster controllers are responsible for installing and managing respective service. Various controllers are installed and managed in a Spark cluster.
32
32
33
-
## Spark service components
33
+
## Apache Spark service components
34
34
35
35
**Zookeeper service:** A three node Zookeeper cluster, serves as distributed coordinator or High Availability storage for other services.
36
36
37
37
**Yarn service:** Hadoop Yarn cluster, Spark jobs would be scheduled in the cluster as Yarn applications.
38
38
39
-
**Client Interfaces:** HDInsight on AKS Spark provides various client interfaces. Livy Server, Jupyter Notebook, Spark History Server, provides Spark services to HDInsight on AKS users.
39
+
**Client Interfaces:** Apache Spark clusters in HDInsight on AKS, provides various client interfaces. Livy Server, Jupyter Notebook, Spark History Server, provides Spark services to HDInsight on AKS users.
40
+
41
+
## Reference
42
+
43
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
@@ -19,13 +19,13 @@ Once the cluster is created, user can use various interfaces to submit and manag
19
19
## Using Jupyter
20
20
21
21
### Prerequisites
22
-
An Apache Spark cluster on HDInsight on AKS. For more information, see [Create an Apache Spark cluster](./create-spark-cluster.md).
22
+
An Apache Spark™ cluster on HDInsight on AKS. For more information, see [Create an Apache Spark cluster](./create-spark-cluster.md).
23
23
24
24
Jupyter Notebook is an interactive notebook environment that supports various programming languages.
25
25
26
26
### Create a Jupyter Notebook
27
27
28
-
1. Navigate to the Spark cluster page and open the **Overview** tab. Click on Jupyter, it asks you to authenticate and open the Jupyter web page.
28
+
1. Navigate to the Apache Spark™ cluster page and open the **Overview** tab. Click on Jupyter, it asks you to authenticate and open the Jupyter web page.
29
29
30
30
:::image type="content" source="./media/submit-manage-jobs/select-jupyter-notebook.png" alt-text="Screenshot of how to select Jupyter notebook." border="true" lightbox="./media/submit-manage-jobs/select-jupyter-notebook.png":::
31
31
@@ -106,13 +106,13 @@ Jupyter Notebook is an interactive notebook environment that supports various pr
106
106
107
107
## Using Apache Zeppelin notebooks
108
108
109
-
HDInsight on AKS Spark clusters include [Apache Zeppelin notebooks](https://zeppelin.apache.org/). Use the notebooks to run Apache Spark jobs. In this article, you learn how to use the Zeppelin notebook on an HDInsight on AKS cluster.
109
+
Apache Spark clusters in HDInsight on AKS include [Apache Zeppelin notebooks](https://zeppelin.apache.org/). Use the notebooks to run Apache Spark jobs. In this article, you learn how to use the Zeppelin notebook on an HDInsight on AKS cluster.
110
110
### Prerequisites
111
111
An Apache Spark cluster on HDInsight on AKS. For instructions, see [Create an Apache Spark cluster](./create-spark-cluster.md).
112
112
113
113
#### Launch an Apache Zeppelin notebook
114
114
115
-
1. Navigate to the Spark cluster Overview page and select Zeppelin notebook from Cluster dashboards. It prompts to authenticate and open the Zeppelin page.
115
+
1. Navigate to the Apache Spark cluster Overview page and select Zeppelin notebook from Cluster dashboards. It prompts to authenticate and open the Zeppelin page.
116
116
117
117
:::image type="content" source="./media/submit-manage-jobs/select-zeppelin.png" alt-text="Screenshot of how to select Zeppelin." lightbox="./media/submit-manage-jobs/select-zeppelin.png":::
118
118
@@ -227,7 +227,7 @@ An Apache Spark cluster on HDInsight on AKS. For instructions, see [Create an
227
227
228
228
:::image type="content" source="./media/submit-manage-jobs/run-spark-submit-job.png" alt-text="Screenshot showing how to run Spark submit job." lightbox="./media/submit-manage-jobs/view-vim-file.png":::
229
229
230
-
## Monitor queries on a Spark cluster in HDInsight on AKS
230
+
## Monitor queries on an Apache Spark cluster in HDInsight on AKS
231
231
232
232
#### Spark History UI
233
233
@@ -264,4 +264,6 @@ An Apache Spark cluster on HDInsight on AKS. For instructions, see [Create an
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
@@ -16,7 +16,7 @@ Azure HDInsight on AKS supports custom meta stores, which are recommended for pr
16
16
17
17
1. Create Azure SQL database
18
18
1. Create a key vault for storing the credentials
19
-
1. Configure Metastore while you create a HDInsight Spark cluster
19
+
1. Configure Metastore while you create a HDInsight on AKS cluster with Apache Spark™
20
20
1. Operate on External Metastore (Shows databases and do a select limit 1).
21
21
22
22
While you create the cluster, HDInsight service needs to connect to the external metastore and verify your credentials.
@@ -68,7 +68,7 @@ While you create the cluster, HDInsight service needs to connect to the external
68
68
69
69
:::image type="content" source="./media/use-hive-metastore/basic-tab.png" alt-text="Screenshot showing the basic tab." lightbox="./media/use-hive-metastore/basic-tab.png":::
70
70
71
-
1. The rest of the details are to be filled in as per the cluster creation rules for [HDInsight on AKS Spark cluster](./create-spark-cluster.md).
71
+
1. The rest of the details are to be filled in as per the cluster creation rules for [Apache Spark cluster in HDInsight on AKS](./create-spark-cluster.md).
72
72
73
73
1. Click on **Review and Create.**
74
74
@@ -97,5 +97,8 @@ While you create the cluster, HDInsight service needs to connect to the external
97
97
`>> spark.sql("select * from sampleTable").show()`
98
98
99
99
:::image type="content" source="./media/use-hive-metastore/read-table.png" alt-text="Screenshot showing how to read table." lightbox="./media/use-hive-metastore/read-table.png":::
100
+
101
+
## Reference
100
102
103
+
* Apache, Apache Spark, Spark, and associated open source project names are [trademarks](../trademarks.md) of the [Apache Software Foundation](https://www.apache.org/) (ASF).
0 commit comments