Skip to content

Commit 7aca269

Browse files
authored
Merge pull request #313754 from lootle1/FR43
Freshness Edit: Azure Dev Compute
2 parents 10d0269 + 97d0fa8 commit 7aca269

1 file changed

Lines changed: 20 additions & 21 deletions

File tree

articles/api-management/azure-ai-foundry-api.md

Lines changed: 20 additions & 21 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ ms.service: azure-api-management
55
author: dlepow
66
ms.author: danlep
77
ms.topic: how-to
8-
ms.date: 10/07/2025
8+
ms.date: 03/24/2026
99
ms.update-cycle: 180-days
1010
ms.collection: ce-skilling-ai-copilot
1111
ms.custom: template-how-to, build-2024
@@ -21,32 +21,31 @@ Learn more about managing AI APIs in API Management:
2121

2222
* [AI gateway capabilities in Azure API Management](genai-gateway-capabilities.md)
2323

24-
2524
## Client compatibility options
2625

2726
API Management supports two client compatibility options for AI APIs from Microsoft Foundry. When you import the API using the wizard, choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the Foundry tool.
2827

29-
* **Azure OpenAI** - Manage Azure OpenAI in Microsoft Foundry model deployments.
28+
* **Azure OpenAI**: Manage Azure OpenAI in Microsoft Foundry model deployments.
3029

31-
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your Foundry tool only includes Azure OpenAI model deployments.
30+
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your Foundry tool only includes Azure OpenAI model deployments.
3231

33-
* **Azure AI** - Manage model endpoints in Microsoft Foundry that are exposed through the [Azure AI Model Inference API](/azure/ai-studio/reference/reference-model-inference-api).
32+
* **Azure AI**: Manage model endpoints in Microsoft Foundry that are exposed through the [Azure AI Model Inference API](/rest/api/aifoundry/modelinference/).
3433

3534
Clients call the deployment at a `/models` endpoint such as `/my-model/models/chat/completions`. Deployment name is passed in the request body. Use this option if you want flexibility to switch between models exposed through the Azure AI Model Inference API and those deployed in Azure OpenAI in Foundry Models.
3635

3736
## Prerequisites
3837

39-
- An existing API Management instance. [Create one if you haven't already](get-started-create-service-instance.md).
38+
* An existing API Management instance. [Create one if you haven't already](get-started-create-service-instance.md).
4039

41-
- A Foundry tool in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
40+
* A Foundry tool in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
4241

4342
## Import Microsoft Foundry API using the portal
4443

45-
Use the following steps to import an AI API to API Management.
44+
Use the following steps to import an AI API to API Management.
4645

4746
When you import the API, API Management automatically configures:
4847

49-
* Operations for each of the API's REST API endpoints
48+
* Operations for each of the API's REST API endpoints.
5049
* A system-assigned identity with the necessary permissions to access the Foundry tool deployment.
5150
* A [backend](backends.md) resource and a [set-backend-service](set-backend-service-policy.md) policy that direct API requests to the Azure AI Services endpoint.
5251
* Authentication to the backend using the instance's system-assigned managed identity.
@@ -62,37 +61,38 @@ To import a Microsoft Foundry API to API Management:
6261
1. On the **Select AI Service** tab:
6362
1. Select the **Subscription** in which to search for Foundry Tools. To get information about the model deployments in a service, select the **deployments** link next to the service name.
6463
:::image type="content" source="media/azure-ai-foundry-api/deployments.png" alt-text="Screenshot of deployments for an AI service in the portal.":::
65-
1. Select a Foundry tool.
64+
1. Select a Foundry tool.
6665
1. Select **Next**.
6766
1. On the **Configure API** tab:
6867
1. Enter a **Display name** and optional **Description** for the API.
6968
1. In **Base path**, enter a path that your API Management instance uses to access the deployment endpoint.
70-
1. Optionally select one or more **Products** to associate with the API.
69+
1. Optionally, select one or more **Products** to associate with the API.
7170
1. In **Client compatibility**, select either of the following based on the types of client you intend to support. See [Client compatibility options](#client-compatibility-options) for more information.
72-
* **Azure OpenAI** - Select this option if your clients only need to access Azure OpenAI in Microsoft Foundry model deployments.
73-
* **Azure AI** - Select this option if your clients need to access other models in Microsoft Foundry.
71+
* **Azure OpenAI**: Select this option if your clients only need to access Azure OpenAI in Microsoft Foundry model deployments.
72+
* **Azure AI**: Select this option if your clients need to access other models in Microsoft Foundry.
7473
1. Select **Next**.
7574

7675
:::image type="content" source="media/azure-ai-foundry-api/client-compatibility.png" alt-text="Screenshot of Microsoft Foundry API configuration in the portal.":::
7776

78-
1. On the **Manage token consumption** tab, optionally enter settings or accept defaults that define the following policies to help monitor and manage the API:
77+
1. On the **Manage token consumption** tab, optionally enter settings, or accept defaults that define the following policies to help monitor and manage the API:
7978
* [Manage token consumption](llm-token-limit-policy.md)
80-
* [Track token usage](llm-emit-token-metric-policy.md)
81-
1. On the **Apply semantic caching** tab, optionally enter settings or accept defaults that define the policies to help optimize performance and reduce latency for the API:
79+
* [Track token usage](llm-emit-token-metric-policy.md)
80+
1. On the **Apply semantic caching** tab, optionally enter settings, or accept defaults that define the policies to help optimize performance and reduce latency for the API:
8281
* [Enable semantic caching of responses](azure-openai-enable-semantic-caching.md)
83-
1. On the **AI content safety**, optionally enter settings or accept defaults to configure the Azure AI Content Safety service to block prompts with unsafe content:
82+
1. On the **AI content safety**, optionally enter settings, or accept defaults to configure the Azure AI Content Safety service to block prompts with unsafe content:
8483
* [Enforce content safety checks on LLM requests](llm-content-safety-policy.md)
8584
1. Select **Review**.
86-
1. After settings are validated, select **Create**.
85+
1. After settings are validated, select **Create**.
8786

8887
## Test the AI API
8988

90-
To ensure that your AI API is working as expected, test it in the API Management test console.
89+
To ensure that your AI API is working as expected, test it in the API Management test console.
90+
9191
1. Select the API you created in the previous step.
9292
1. Select the **Test** tab.
9393
1. Select an operation that's compatible with the model deployment.
9494
The page displays fields for parameters and headers.
95-
1. Enter parameters and headers as needed. Depending on the operation, you might need to configure or update a **Request body**. Here's a very basic example request body for a chat completions operation:
95+
1. Enter parameters and headers as needed. Depending on the operation, you might need to configure or update a **Request body**. Here's a basic example request body for a chat completions operation:
9696

9797
```json
9898
{
@@ -113,5 +113,4 @@ To ensure that your AI API is working as expected, test it in the API Management
113113

114114
When the test is successful, the backend responds with a successful HTTP response code and some data. Appended to the response is token usage data to help you monitor and manage your language model token consumption.
115115

116-
117116
[!INCLUDE [api-management-define-api-topics.md](../../includes/api-management-define-api-topics.md)]

0 commit comments

Comments
 (0)