Skip to content

Commit 67666b7

Browse files
committed
Update Azure AI Foundry and Azure AI services references
1 parent 454258d commit 67666b7

33 files changed

Lines changed: 69 additions & 69 deletions

articles/api-management/amazon-bedrock-passthrough-llm-api.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ ms.custom: template-how-to, build-2024
1515

1616
[!INCLUDE [api-management-availability-all-tiers](../../includes/api-management-availability-all-tiers.md)]
1717

18-
In this article, you import an Amazon Bedrock language model API into your API Management instance as a passthrough API. This is an example of a model that's hosted on an inference provider other than Azure AI services. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
18+
In this article, you import an Amazon Bedrock language model API into your API Management instance as a passthrough API. This is an example of a model that's hosted on an inference provider other than Foundry Tools. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
1919

2020
Learn more about managing AI APIs in API Management:
2121

articles/api-management/api-management-authenticate-authorize-ai-apis.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -102,7 +102,7 @@ Follow these steps to configure your API Management instance to use a managed id
102102
```
103103

104104
> [!TIP]
105-
> Instead of using the `authentication-managed-identity` and `set-header` policies shown in this example, you can configure a [backend](backends.md) resource that directs API requests to the AI service endpoint. In the backend configuration, configure managed identity credentials to the `https://cognitiveservices.azure.com/` resource. Azure API Management automates these steps when you [import an API directly from Microsoft Foundry](azure-ai-foundry-api.md).
105+
> Instead of using the `authentication-managed-identity` and `set-header` policies shown in this example, you can configure a [backend](backends.md) resource that directs API requests to the Azure AI Services endpoint. In the backend configuration, configure managed identity credentials to the `https://cognitiveservices.azure.com/` resource. Azure API Management automates these steps when you [import an API directly from Microsoft Foundry](azure-ai-foundry-api.md).
106106

107107
## OAuth 2.0 authorization by using identity provider
108108

articles/api-management/azure-ai-foundry-api.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -24,11 +24,11 @@ Learn more about managing AI APIs in API Management:
2424

2525
## Client compatibility options
2626

27-
API Management supports two client compatibility options for AI APIs from Microsoft Foundry. When you import the API using the wizard, choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the AI service.
27+
API Management supports two client compatibility options for AI APIs from Microsoft Foundry. When you import the API using the wizard, choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the Foundry Tool.
2828

2929
* **Azure OpenAI** - Manage Azure OpenAI in Microsoft Foundry model deployments.
3030

31-
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your AI service only includes Azure OpenAI model deployments.
31+
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your Foundry Tool only includes Azure OpenAI model deployments.
3232

3333
* **Azure AI** - Manage model endpoints in Microsoft Foundry that are exposed through the [Azure AI Model Inference API](/azure/ai-studio/reference/reference-model-inference-api).
3434

@@ -38,7 +38,7 @@ API Management supports two client compatibility options for AI APIs from Micros
3838

3939
- An existing API Management instance. [Create one if you haven't already](get-started-create-service-instance.md).
4040

41-
- An Azure AI service in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
41+
- A Foundry Tool in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
4242

4343
## Import Microsoft Foundry API using the portal
4444

@@ -47,8 +47,8 @@ Use the following steps to import an AI API to API Management.
4747
When you import the API, API Management automatically configures:
4848

4949
* Operations for each of the API's REST API endpoints
50-
* A system-assigned identity with the necessary permissions to access the AI service deployment.
51-
* A [backend](backends.md) resource and a [set-backend-service](set-backend-service-policy.md) policy that direct API requests to the AI service endpoint.
50+
* A system-assigned identity with the necessary permissions to access the Foundry Tool deployment.
51+
* A [backend](backends.md) resource and a [set-backend-service](set-backend-service-policy.md) policy that direct API requests to the Azure AI Services endpoint.
5252
* Authentication to the backend using the instance's system-assigned managed identity.
5353
* (optionally) Policies to help you monitor and manage the API.
5454

@@ -59,10 +59,10 @@ To import a Microsoft Foundry API to API Management:
5959
1. Under **Create from Azure resource**, select **Microsoft Foundry**.
6060

6161
:::image type="content" source="media/azure-ai-foundry-api/ai-foundry-api.png" alt-text="Screenshot of creating an OpenAI-compatible API in the portal." :::
62-
1. On the **Select AI service** tab:
63-
1. Select the **Subscription** in which to search for AI services. To get information about the model deployments in a service, select the **deployments** link next to the service name.
62+
1. On the **Select AI Service** tab:
63+
1. Select the **Subscription** in which to search for Foundry Tools. To get information about the model deployments in a service, select the **deployments** link next to the service name.
6464
:::image type="content" source="media/azure-ai-foundry-api/deployments.png" alt-text="Screenshot of deployments for an AI service in the portal.":::
65-
1. Select an AI service.
65+
1. Select a Foundry Tool.
6666
1. Select **Next**.
6767
1. On the **Configure API** tab:
6868
1. Enter a **Display name** and optional **Description** for the API.

articles/api-management/backends.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ For other APIs, such as APIs from Azure services, you import an Azure resource w
2828

2929
API Management also supports using other resources as an API backend, such as:
3030
* A [Service Fabric cluster](how-to-configure-service-fabric-backend.yml).
31-
* AI services.
31+
* Foundry Tools.
3232
* A custom service.
3333

3434
For these backends, you can create a *backend entity* in API Management and reference it in your APIs.
@@ -51,7 +51,7 @@ You can configure and manage backend entities in the Azure portal, or by using A
5151
You can create a backend in the Azure portal, or by using Azure APIs or tools.
5252

5353
> [!NOTE]
54-
> When you import certain APIs, such as APIs from Microsoft Foundry or other AI services, API Management automatically configures a backend entity.
54+
> When you import certain APIs, such as APIs from Microsoft Foundry or other Foundry Tools, API Management automatically configures a backend entity.
5555
5656
To create a backend in the portal:
5757

articles/api-management/breaking-changes/identity-provider-adal-retirement-sep-2025.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ This change will have no effect on the availability of your API Management servi
2323

2424
Your service is impacted by this change if:
2525

26-
* You've configured an [Microsoft Entra ID](../api-management-howto-aad.md) or [Azure AD B2C](../api-management-howto-aad-b2c.md) identity provider for user account authentication using the ADAL and use the provided developer portal.
26+
* You've configured a [Microsoft Entra ID](../api-management-howto-aad.md) or [Azure AD B2C](../api-management-howto-aad-b2c.md) identity provider for user account authentication using the ADAL and use the provided developer portal.
2727

2828
## What is the deadline for the change?
2929

articles/api-management/edit-api.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Edit an API in the Azure Portal | Microsoft Docs
2+
title: Edit an API in the Azure portal | Microsoft Docs
33
description: Learn how to use API Management to edit an API or its swagger.
44
services: api-management
55
author: dlepow

articles/api-management/front-door-api-management.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -143,7 +143,7 @@ The following are high level steps to add an endpoint for the developer portal t
143143
For more information and details about settings, see [How to configure an origin for Azure Front Door](../frontdoor/how-to-configure-origin.md#create-a-new-origin-group).
144144

145145
> [!NOTE]
146-
> If you've configured an [Microsoft Entra ID](api-management-howto-aad.md) or [Microsoft Entra External ID](/entra/external-id/customers/overview-customers-ciam) identity provider for the developer portal, you need to update the corresponding app registration with an additional redirect URL to Front Door. In the app registration, add the URL for the developer portal endpoint configured in your Front Door profile.
146+
> If you've configured a [Microsoft Entra ID](api-management-howto-aad.md) or [Microsoft Entra External ID](/entra/external-id/customers/overview-customers-ciam) identity provider for the developer portal, you need to update the corresponding app registration with an additional redirect URL to Front Door. In the app registration, add the URL for the developer portal endpoint configured in your Front Door profile.
147147
148148
## Related content
149149

articles/api-management/genai-gateway-capabilities.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ AI adoption in organizations involves several phases:
4545

4646
As AI adoption matures, especially in larger enterprises, the AI gateway helps address key challenges. It helps you:
4747

48-
* Authenticate and authorize access to AI services
48+
* Authenticate and authorize access to Foundry Tools
4949
* Load balance across multiple AI endpoints
5050
* Monitor and log AI interactions
5151
* Manage token usage and quotas across multiple applications
@@ -68,7 +68,7 @@ For example, to onboard a model deployed in Microsoft Foundry or another provide
6868

6969
More information:
7070

71-
* [Import an Microsoft Foundry API](azure-ai-foundry-api.md)
71+
* [Import a Microsoft Foundry API](azure-ai-foundry-api.md)
7272
* [Import a language model API](openai-compatible-llm-api.md)
7373
* [Expose a REST API as an MCP server](export-rest-mcp-server.md)
7474
* [Expose and govern an existing MCP server](expose-existing-mcp-server.md)
@@ -78,15 +78,15 @@ More information:
7878

7979
One of the main resources in generative AI services is *tokens*. Microsoft Foundry and other providers assign quotas for your model deployments as tokens per minute (TPM). You distribute these tokens across your model consumers, such as different applications, developer teams, or departments within the company.
8080

81-
If you have a single app connecting to an AI service backend, you can manage token consumption with a TPM limit that you set directly on the model deployment. However, when your application portfolio grows, you might have multiple apps calling single or multiple AI service endpoints. These endpoints can be pay-as-you-go or [Provisioned Throughput Units](/azure/ai-services/openai/concepts/provisioned-throughput) (PTU) instances. You need to make sure that one app doesn't use the whole TPM quota and block other apps from accessing the backends they need.
81+
If you have a single app connecting to an AI service backend, you can manage token consumption with a TPM limit that you set directly on the model deployment. However, when your application portfolio grows, you might have multiple apps calling single or multiple Azure AI Services endpoints. These endpoints can be pay-as-you-go or [Provisioned Throughput Units](/azure/ai-services/openai/concepts/provisioned-throughput) (PTU) instances. You need to make sure that one app doesn't use the whole TPM quota and block other apps from accessing the backends they need.
8282

8383
### Token rate limiting and quotas
8484

85-
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of AI service tokens. By using this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
85+
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of Foundry Tool tokens. By using this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
8686

8787
:::image type="content" source="media/genai-gateway-capabilities/token-rate-limiting.png" alt-text="Diagram of limiting Azure OpenAI Service tokens in API Management.":::
8888

89-
This policy provides flexibility to assign token-based limits on any counter key, such as subscription key, originating IP address, or an arbitrary key defined through a policy expression. The policy also enables precalculation of prompt tokens on the Azure API Management side, minimizing unnecessary requests to the AI service backend if the prompt already exceeds the limit.
89+
This policy provides flexibility to assign token-based limits on any counter key, such as subscription key, originating IP address, or an arbitrary key defined through a policy expression. The policy also enables precalculation of prompt tokens on the Azure API Management side, minimizing unnecessary requests to the Foundry Tool backend if the prompt already exceeds the limit.
9090

9191
The following basic example demonstrates how to set a TPM limit of 500 per subscription key:
9292

@@ -102,7 +102,7 @@ More information:
102102

103103
### Semantic caching
104104

105-
Semantic caching is a technique that improves the performance of LLM APIs by caching the results (completions) of previous prompts and reusing them by comparing the vector proximity of the prompt to prior requests. This technique reduces the number of calls made to the AI service backend, improves response times for end users, and can help reduce costs.
105+
Semantic caching is a technique that improves the performance of LLM APIs by caching the results (completions) of previous prompts and reusing them by comparing the vector proximity of the prompt to prior requests. This technique reduces the number of calls made to the Foundry Tool backend, improves response times for end users, and can help reduce costs.
106106

107107
In API Management, enable semantic caching by using [Azure Managed Redis](/azure/redis/overview) or another external cache compatible with RediSearch and onboarded to Azure API Management. By using the Embeddings API, the [llm-semantic-cache-store](llm-semantic-cache-store-policy.md) and [llm-semantic-cache-lookup](llm-semantic-cache-lookup-policy.md) policies store and retrieve semantically similar prompt completions from the cache. This approach ensures completions reuse, resulting in reduced token consumption and improved response performance.
108108

@@ -124,13 +124,13 @@ More information:
124124
* [Deploy an API Management instance in multiple regions](api-management-howto-deploy-multi-region.md)
125125

126126
> [!NOTE]
127-
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, deploy backend AI services in the same regions as your API Management gateways.
127+
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, deploy backend Foundry Tools in the same regions as your API Management gateways.
128128
129129
## Security and safety
130130

131131
An AI gateway secures and controls access to your AI APIs. By using the AI gateway, you can:
132132

133-
* Use managed identities to authenticate to Azure AI services, so you don't need API keys for authentication
133+
* Use managed identities to authenticate to Foundry Tools, so you don't need API keys for authentication
134134
* Configure OAuth authorization for AI apps and agents to access APIs or MCP servers by using API Management's credential manager
135135
* Apply policies to automatically moderate LLM prompts by using [Azure AI Content Safety](/azure/ai-services/content-safety/overview)
136136

@@ -146,7 +146,7 @@ More information:
146146

147147
## Resiliency
148148

149-
One challenge when building intelligent applications is ensuring that the applications are resilient to backend failures and can handle high loads. By configuring your LLM endpoints with [backends](backends.md) in Azure API Management, you can balance the load across them. You can also define circuit breaker rules to stop forwarding requests to AI service backends if they're not responsive.
149+
One challenge when building intelligent applications is ensuring that the applications are resilient to backend failures and can handle high loads. By configuring your LLM endpoints with [backends](backends.md) in Azure API Management, you can balance the load across them. You can also define circuit breaker rules to stop forwarding requests to Foundry Tool backends if they're not responsive.
150150

151151
### Load balancer
152152

articles/api-management/openai-compatible-llm-api.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,15 +15,15 @@ ms.custom: template-how-to
1515

1616
[!INCLUDE [api-management-availability-all-tiers](../../includes/api-management-availability-all-tiers.md)]
1717

18-
You can import OpenAI-compatible language model endpoints to your API Management instance as APIs. You can also import language models that aren't compatible with OpenAI as passthrough APIs, which forward requests directly to the backend endpoints. For example, you might want to manage an LLM that you self-host, or that's hosted on an inference provider other than Azure AI services. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
18+
You can import OpenAI-compatible language model endpoints to your API Management instance as APIs. You can also import language models that aren't compatible with OpenAI as passthrough APIs, which forward requests directly to the backend endpoints. For example, you might want to manage an LLM that you self-host, or that's hosted on an inference provider other than Foundry Tools. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
1919

2020
Learn more about managing AI APIs in API Management:
2121

2222
* [AI gateway capabilities in Azure API Management](genai-gateway-capabilities.md)
2323

2424
## Language model API types
2525

26-
API Management supports two types of language model APIs for this scenario. Choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the AI service.
26+
API Management supports two types of language model APIs for this scenario. Choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the Foundry Tool.
2727

2828
* **OpenAI-compatible** - Language model endpoints that are compatible with OpenAI's API. Examples include certain models exposed by inference providers such as [Hugging Face Text Generation Inference (TGI)](https://huggingface.co/docs/text-generation-inference/en/index) and [Google Gemini API](openai-compatible-google-gemini-api.md).
2929

articles/api-management/sap-api.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
---
2-
title: Import an SAP API by Using the Azure Portal | Microsoft Docs
2+
title: Import an SAP API by Using the Azure portal | Microsoft Docs
33
titleSuffix:
44
description: Learn how to import OData metadata from SAP as an API to Azure API Management, either directly or by converting the metadata to an OpenAPI specification.
55
ms.service: azure-api-management

0 commit comments

Comments
 (0)