Skip to content

Commit 9ac8f83

Browse files
committed
Capture Event Hubs - Parquet
1 parent cc9eb36 commit 9ac8f83

1 file changed

Lines changed: 13 additions & 11 deletions

File tree

articles/stream-analytics/capture-event-hub-data-parquet.md

Lines changed: 13 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,12 @@
11
---
2-
title: Event Hubs to Azure Data Lake in Parquet format
2+
title: Event Hubs Data Capture to Azure Data Lake Parquet
33
description: Learn how to use the node code editor to automatically capture the streaming data in Event Hubs in an Azure Data Lake Storage Gen2 account in Parquet format.
44
author: xujxu
55
ms.author: xujiang1
6+
ms.reviewer: spelluru
67
ms.service: azure-stream-analytics
78
ms.topic: how-to
8-
ms.date: 01/23/2025
9+
ms.date: 03/26/2026
910
ms.custom:
1011
- mvc
1112
- sfi-image-nochange
@@ -20,7 +21,7 @@ This article explains how to use the no code editor to automatically capture str
2021

2122
If you don't have an event hub, create one by following instructions from [Quickstart: Create an event hub](../event-hubs/event-hubs-create.md).
2223

23-
If you don't have a Data Lake Storage Gen2 account, create one by following instructions from [Create a storage account](../storage/blobs/create-data-lake-storage-account.md)
24+
If you don't have a Data Lake Storage Gen2 account, create one by following instructions from [Create a storage account](../storage/blobs/create-data-lake-storage-account.md).
2425
- The data in your Event Hubs instance (event hub) must be serialized in either JSON, CSV, or Avro format. On the **Event Hubs Instance** page for your event hub, follow these steps:
2526
1. On the left menu, select **Data Explorer**.
2627
1. In the middle pane, select **Send events**.
@@ -33,8 +34,8 @@ This article explains how to use the no code editor to automatically capture str
3334

3435
Use the following steps to configure a Stream Analytics job to capture data in Azure Data Lake Storage Gen2.
3536

36-
1. In the Azure portal, navigate to your event hub.
37-
1. On the left menu, select **Process Data** under **Features**. Then, select **Start** on the **Capture data to ADLS Gen2 in Parquet format** card.
37+
1. In the Azure portal, go to your event hub.
38+
1. On the left menu, under **Features**, select **Process Data**. Then, select **Start** on the **Capture data to ADLS Gen2 in Parquet format** card.
3839

3940
:::image type="content" source="./media/capture-event-hub-data-parquet/process-event-hub-data-cards.png" alt-text="Screenshot showing the Process Event Hubs data start cards." lightbox="./media/capture-event-hub-data-parquet/process-event-hub-data-cards.png" :::
4041
1. Enter a **name** for your Stream Analytics job, and then select **Create**.
@@ -51,25 +52,26 @@ Use the following steps to configure a Stream Analytics job to capture data in A
5152
1. Select the **Azure Data Lake Storage Gen2** tile to edit the configuration.
5253
1. On the **Azure Data Lake Storage Gen2** configuration page, follow these steps:
5354
1. Select the subscription, storage account name, and container from the drop-down menu.
54-
1. Once the subscription is selected, the authentication method and storage account key should be automatically filled in.
55+
1. After you select the subscription, the authentication method and storage account key are automatically filled in.
5556
1. Select **Parquet** for **Serialization** format.
5657

5758
:::image type="content" source="./media/capture-event-hub-data-parquet/job-top-settings.png" alt-text="Screenshot showing the Data Lake Storage Gen2 configuration page." lightbox="./media/capture-event-hub-data-parquet/job-top-settings.png":::
58-
1. For streaming blobs, the directory path pattern is expected to be a dynamic value. It's required for the date to be a part of the file path for the blob – referenced as `{date}`. To learn about custom path patterns, see to [Azure Stream Analytics custom blob output partitioning](stream-analytics-custom-path-patterns-blob-storage-output.md).
59+
1. For streaming blobs, the directory path pattern is a dynamic value. The date must be part of the file path for the blob – referenced as `{date}`. To learn about custom path patterns, see [Azure Stream Analytics custom blob output partitioning](stream-analytics-custom-path-patterns-blob-storage-output.md).
5960

6061
:::image type="content" source="./media/capture-event-hub-data-parquet/blob-configuration.png" alt-text="First screenshot showing the Blob window where you edit a blob's connection configuration." lightbox="./media/capture-event-hub-data-parquet/blob-configuration.png" :::
6162
1. Select **Connect**
6263
1. When the connection is established, you see fields that are present in the output data.
6364
1. Select **Save** on the command bar to save your configuration.
6465

65-
:::image type="content" source="./media/capture-event-hub-data-parquet/save-configuration.png" alt-text="Screenshot showing the Save button selected on the command bar." :::
66-
1. Select **Start** on the command bar to start the streaming flow to capture data. Then in the Start Stream Analytics job window:
66+
:::image type="content" source="./media/capture-event-hub-data-parquet/save-configuration.png" alt-text="Screenshot showing the Save button on the command bar." :::
67+
1. Select **Start** on the command bar to start the streaming flow to capture data. Then in the **Start Stream Analytics job** window:
6768
1. Choose the output start time.
6869
1. Select the pricing plan.
6970
1. Select the number of Streaming Units (SU) that the job runs with. SU represents the computing resources that are allocated to execute a Stream Analytics job. For more information, see [Streaming Units in Azure Stream Analytics](stream-analytics-streaming-unit-consumption.md).
7071

7172
:::image type="content" source="./media/capture-event-hub-data-parquet/start-job.png" alt-text="Screenshot showing the Start Stream Analytics job window where you set the output start time, streaming units, and error handling." lightbox="./media/capture-event-hub-data-parquet/start-job.png" :::
72-
1. You should see the Stream Analytic job in the **Stream Analytics job** tab of the **Process data** page for your event hub.
73+
1. Select **X** at the top-right corner to close the **Stream Analytics job** window.
74+
1. You see the Stream Analytic job in the **Stream Analytics job** tab of the **Process data** page for your event hub.
7375

7476
:::image type="content" source="./media/capture-event-hub-data-parquet/process-data-page-jobs.png" alt-text="Screenshot showing the Stream Analytics job on the Process data page." lightbox="./media/capture-event-hub-data-parquet/process-data-page-jobs.png" :::
7577

@@ -84,7 +86,7 @@ Use the following steps to configure a Stream Analytics job to capture data in A
8486
1. Verify that the Parquet files are generated in the Azure Data Lake Storage container.
8587

8688
:::image type="content" source="./media/capture-event-hub-data-parquet/verify-captured-data.png" alt-text="Screenshot showing the generated Parquet files in the Azure Data Lake Storage container." lightbox="./media/capture-event-hub-data-parquet/verify-captured-data.png" :::
87-
1. Back on the Event Hubs instance page, select **Process data** on the left menu. Switch to the **Stream Analytics jobs** tab. Select **Open metrics** to monitor it. Add **Input metrics** to the chart using the **Add metric** on the toolbar. If you don't see the metrics in the chart, wait for a few minutes, and refresh the page.
89+
1. Now, on the Event Hubs instance page, select **Process data** in the left menu. Switch to the **Stream Analytics jobs** tab. Select **Open metrics** to monitor it. Add **Input metrics** to the chart using the **Add metric** on the toolbar. If you don't see the metrics in the chart, wait for a few minutes, and refresh the page.
8890

8991
:::image type="content" source="./media/capture-event-hub-data-parquet/open-metrics-link.png" alt-text="Screenshot showing Open Metrics link selected." lightbox="./media/capture-event-hub-data-parquet/open-metrics-link.png" :::
9092

0 commit comments

Comments
 (0)