You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this article, you import an Amazon Bedrock language model API into your API Management instance as a passthrough API. This is an example of a model that's hosted on an inference provider other than Azure AI services. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
18
+
In this article, you import an Amazon Bedrock language model API into your API Management instance as a passthrough API. This is an example of a model that's hosted on an inference provider other than Foundry Tools. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
19
19
20
20
Learn more about managing AI APIs in API Management:
Copy file name to clipboardExpand all lines: articles/api-management/api-management-authenticate-authorize-ai-apis.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -102,7 +102,7 @@ Follow these steps to configure your API Management instance to use a managed id
102
102
```
103
103
104
104
> [!TIP]
105
-
> Instead of using the `authentication-managed-identity` and `set-header` policies shown in this example, you can configure a [backend](backends.md) resource that directs API requests to the AI service endpoint. In the backend configuration, configure managed identity credentials to the `https://cognitiveservices.azure.com/` resource. Azure API Management automates these steps when you [import an API directly from Microsoft Foundry](azure-ai-foundry-api.md).
105
+
> Instead of using the `authentication-managed-identity` and `set-header` policies shown in this example, you can configure a [backend](backends.md) resource that directs API requests to the Azure AI Services endpoint. In the backend configuration, configure managed identity credentials to the `https://cognitiveservices.azure.com/` resource. Azure API Management automates these steps when you [import an API directly from Microsoft Foundry](azure-ai-foundry-api.md).
106
106
107
107
## OAuth 2.0 authorization by using identity provider
Copy file name to clipboardExpand all lines: articles/api-management/azure-ai-foundry-api.md
+8-8Lines changed: 8 additions & 8 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,11 +24,11 @@ Learn more about managing AI APIs in API Management:
24
24
25
25
## Client compatibility options
26
26
27
-
API Management supports two client compatibility options for AI APIs from Microsoft Foundry. When you import the API using the wizard, choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the AI service.
27
+
API Management supports two client compatibility options for AI APIs from Microsoft Foundry. When you import the API using the wizard, choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the Foundry Tool.
28
28
29
29
***Azure OpenAI** - Manage Azure OpenAI in Microsoft Foundry model deployments.
30
30
31
-
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your AI service only includes Azure OpenAI model deployments.
31
+
Clients call the deployment at an `/openai` endpoint such as `/openai/deployments/my-deployment/chat/completions`. Deployment name is passed in the request path. Use this option if your Foundry Tool only includes Azure OpenAI model deployments.
32
32
33
33
***Azure AI** - Manage model endpoints in Microsoft Foundry that are exposed through the [Azure AI Model Inference API](/azure/ai-studio/reference/reference-model-inference-api).
34
34
@@ -38,7 +38,7 @@ API Management supports two client compatibility options for AI APIs from Micros
38
38
39
39
- An existing API Management instance. [Create one if you haven't already](get-started-create-service-instance.md).
40
40
41
-
-An Azure AI service in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
41
+
-A Foundry Tool in your subscription with one or more models deployed. Examples include models deployed in Microsoft Foundry or Azure OpenAI.
42
42
43
43
## Import Microsoft Foundry API using the portal
44
44
@@ -47,8 +47,8 @@ Use the following steps to import an AI API to API Management.
47
47
When you import the API, API Management automatically configures:
48
48
49
49
* Operations for each of the API's REST API endpoints
50
-
* A system-assigned identity with the necessary permissions to access the AI service deployment.
51
-
* A [backend](backends.md) resource and a [set-backend-service](set-backend-service-policy.md) policy that direct API requests to the AI service endpoint.
50
+
* A system-assigned identity with the necessary permissions to access the Foundry Tool deployment.
51
+
* A [backend](backends.md) resource and a [set-backend-service](set-backend-service-policy.md) policy that direct API requests to the Azure AI Services endpoint.
52
52
* Authentication to the backend using the instance's system-assigned managed identity.
53
53
* (optionally) Policies to help you monitor and manage the API.
54
54
@@ -59,10 +59,10 @@ To import a Microsoft Foundry API to API Management:
59
59
1. Under **Create from Azure resource**, select **Microsoft Foundry**.
60
60
61
61
:::image type="content" source="media/azure-ai-foundry-api/ai-foundry-api.png" alt-text="Screenshot of creating an OpenAI-compatible API in the portal." :::
62
-
1. On the **Select AI service** tab:
63
-
1. Select the **Subscription** in which to search for AI services. To get information about the model deployments in a service, select the **deployments** link next to the service name.
62
+
1. On the **Select AI Service** tab:
63
+
1. Select the **Subscription** in which to search for Foundry Tools. To get information about the model deployments in a service, select the **deployments** link next to the service name.
64
64
:::image type="content" source="media/azure-ai-foundry-api/deployments.png" alt-text="Screenshot of deployments for an AI service in the portal.":::
65
-
1. Select an AI service.
65
+
1. Select a Foundry Tool.
66
66
1. Select **Next**.
67
67
1. On the **Configure API** tab:
68
68
1. Enter a **Display name** and optional **Description** for the API.
Copy file name to clipboardExpand all lines: articles/api-management/breaking-changes/identity-provider-adal-retirement-sep-2025.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -23,7 +23,7 @@ This change will have no effect on the availability of your API Management servi
23
23
24
24
Your service is impacted by this change if:
25
25
26
-
* You've configured an[Microsoft Entra ID](../api-management-howto-aad.md) or [Azure AD B2C](../api-management-howto-aad-b2c.md) identity provider for user account authentication using the ADAL and use the provided developer portal.
26
+
* You've configured a[Microsoft Entra ID](../api-management-howto-aad.md) or [Azure AD B2C](../api-management-howto-aad-b2c.md) identity provider for user account authentication using the ADAL and use the provided developer portal.
Copy file name to clipboardExpand all lines: articles/api-management/front-door-api-management.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -143,7 +143,7 @@ The following are high level steps to add an endpoint for the developer portal t
143
143
For more information and details about settings, see [How to configure an origin for Azure Front Door](../frontdoor/how-to-configure-origin.md#create-a-new-origin-group).
144
144
145
145
> [!NOTE]
146
-
> If you've configured an[Microsoft Entra ID](api-management-howto-aad.md) or [Microsoft Entra External ID](/entra/external-id/customers/overview-customers-ciam) identity provider for the developer portal, you need to update the corresponding app registration with an additional redirect URL to Front Door. In the app registration, add the URL for the developer portal endpoint configured in your Front Door profile.
146
+
> If you've configured a[Microsoft Entra ID](api-management-howto-aad.md) or [Microsoft Entra External ID](/entra/external-id/customers/overview-customers-ciam) identity provider for the developer portal, you need to update the corresponding app registration with an additional redirect URL to Front Door. In the app registration, add the URL for the developer portal endpoint configured in your Front Door profile.
Copy file name to clipboardExpand all lines: articles/api-management/genai-gateway-capabilities.md
+9-9Lines changed: 9 additions & 9 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -45,7 +45,7 @@ AI adoption in organizations involves several phases:
45
45
46
46
As AI adoption matures, especially in larger enterprises, the AI gateway helps address key challenges. It helps you:
47
47
48
-
* Authenticate and authorize access to AI services
48
+
* Authenticate and authorize access to Foundry Tools
49
49
* Load balance across multiple AI endpoints
50
50
* Monitor and log AI interactions
51
51
* Manage token usage and quotas across multiple applications
@@ -68,7 +68,7 @@ For example, to onboard a model deployed in Microsoft Foundry or another provide
68
68
69
69
More information:
70
70
71
-
*[Import an Microsoft Foundry API](azure-ai-foundry-api.md)
71
+
*[Import a Microsoft Foundry API](azure-ai-foundry-api.md)
72
72
*[Import a language model API](openai-compatible-llm-api.md)
73
73
*[Expose a REST API as an MCP server](export-rest-mcp-server.md)
74
74
*[Expose and govern an existing MCP server](expose-existing-mcp-server.md)
@@ -78,15 +78,15 @@ More information:
78
78
79
79
One of the main resources in generative AI services is *tokens*. Microsoft Foundry and other providers assign quotas for your model deployments as tokens per minute (TPM). You distribute these tokens across your model consumers, such as different applications, developer teams, or departments within the company.
80
80
81
-
If you have a single app connecting to an AI service backend, you can manage token consumption with a TPM limit that you set directly on the model deployment. However, when your application portfolio grows, you might have multiple apps calling single or multiple AI service endpoints. These endpoints can be pay-as-you-go or [Provisioned Throughput Units](/azure/ai-services/openai/concepts/provisioned-throughput) (PTU) instances. You need to make sure that one app doesn't use the whole TPM quota and block other apps from accessing the backends they need.
81
+
If you have a single app connecting to an AI service backend, you can manage token consumption with a TPM limit that you set directly on the model deployment. However, when your application portfolio grows, you might have multiple apps calling single or multiple Azure AI Services endpoints. These endpoints can be pay-as-you-go or [Provisioned Throughput Units](/azure/ai-services/openai/concepts/provisioned-throughput) (PTU) instances. You need to make sure that one app doesn't use the whole TPM quota and block other apps from accessing the backends they need.
82
82
83
83
### Token rate limiting and quotas
84
84
85
-
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of AI service tokens. By using this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
85
+
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of Foundry Tool tokens. By using this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
86
86
87
87
:::image type="content" source="media/genai-gateway-capabilities/token-rate-limiting.png" alt-text="Diagram of limiting Azure OpenAI Service tokens in API Management.":::
88
88
89
-
This policy provides flexibility to assign token-based limits on any counter key, such as subscription key, originating IP address, or an arbitrary key defined through a policy expression. The policy also enables precalculation of prompt tokens on the Azure API Management side, minimizing unnecessary requests to the AI service backend if the prompt already exceeds the limit.
89
+
This policy provides flexibility to assign token-based limits on any counter key, such as subscription key, originating IP address, or an arbitrary key defined through a policy expression. The policy also enables precalculation of prompt tokens on the Azure API Management side, minimizing unnecessary requests to the Foundry Tool backend if the prompt already exceeds the limit.
90
90
91
91
The following basic example demonstrates how to set a TPM limit of 500 per subscription key:
92
92
@@ -102,7 +102,7 @@ More information:
102
102
103
103
### Semantic caching
104
104
105
-
Semantic caching is a technique that improves the performance of LLM APIs by caching the results (completions) of previous prompts and reusing them by comparing the vector proximity of the prompt to prior requests. This technique reduces the number of calls made to the AI service backend, improves response times for end users, and can help reduce costs.
105
+
Semantic caching is a technique that improves the performance of LLM APIs by caching the results (completions) of previous prompts and reusing them by comparing the vector proximity of the prompt to prior requests. This technique reduces the number of calls made to the Foundry Tool backend, improves response times for end users, and can help reduce costs.
106
106
107
107
In API Management, enable semantic caching by using [Azure Managed Redis](/azure/redis/overview) or another external cache compatible with RediSearch and onboarded to Azure API Management. By using the Embeddings API, the [llm-semantic-cache-store](llm-semantic-cache-store-policy.md) and [llm-semantic-cache-lookup](llm-semantic-cache-lookup-policy.md) policies store and retrieve semantically similar prompt completions from the cache. This approach ensures completions reuse, resulting in reduced token consumption and improved response performance.
108
108
@@ -124,13 +124,13 @@ More information:
124
124
*[Deploy an API Management instance in multiple regions](api-management-howto-deploy-multi-region.md)
125
125
126
126
> [!NOTE]
127
-
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, deploy backend AI services in the same regions as your API Management gateways.
127
+
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, deploy backend Foundry Tools in the same regions as your API Management gateways.
128
128
129
129
## Security and safety
130
130
131
131
An AI gateway secures and controls access to your AI APIs. By using the AI gateway, you can:
132
132
133
-
* Use managed identities to authenticate to Azure AI services, so you don't need API keys for authentication
133
+
* Use managed identities to authenticate to Foundry Tools, so you don't need API keys for authentication
134
134
* Configure OAuth authorization for AI apps and agents to access APIs or MCP servers by using API Management's credential manager
135
135
* Apply policies to automatically moderate LLM prompts by using [Azure AI Content Safety](/azure/ai-services/content-safety/overview)
136
136
@@ -146,7 +146,7 @@ More information:
146
146
147
147
## Resiliency
148
148
149
-
One challenge when building intelligent applications is ensuring that the applications are resilient to backend failures and can handle high loads. By configuring your LLM endpoints with [backends](backends.md) in Azure API Management, you can balance the load across them. You can also define circuit breaker rules to stop forwarding requests to AI service backends if they're not responsive.
149
+
One challenge when building intelligent applications is ensuring that the applications are resilient to backend failures and can handle high loads. By configuring your LLM endpoints with [backends](backends.md) in Azure API Management, you can balance the load across them. You can also define circuit breaker rules to stop forwarding requests to Foundry Tool backends if they're not responsive.
You can import OpenAI-compatible language model endpoints to your API Management instance as APIs. You can also import language models that aren't compatible with OpenAI as passthrough APIs, which forward requests directly to the backend endpoints. For example, you might want to manage an LLM that you self-host, or that's hosted on an inference provider other than Azure AI services. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
18
+
You can import OpenAI-compatible language model endpoints to your API Management instance as APIs. You can also import language models that aren't compatible with OpenAI as passthrough APIs, which forward requests directly to the backend endpoints. For example, you might want to manage an LLM that you self-host, or that's hosted on an inference provider other than Foundry Tools. Use AI gateway policies and other capabilities in API Management to simplify integration, improve observability, and enhance control over the model endpoints.
19
19
20
20
Learn more about managing AI APIs in API Management:
21
21
22
22
*[AI gateway capabilities in Azure API Management](genai-gateway-capabilities.md)
23
23
24
24
## Language model API types
25
25
26
-
API Management supports two types of language model APIs for this scenario. Choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the AI service.
26
+
API Management supports two types of language model APIs for this scenario. Choose the option suitable for your model deployment. The option determines how clients call the API and how the API Management instance routes requests to the Foundry Tool.
27
27
28
28
***OpenAI-compatible** - Language model endpoints that are compatible with OpenAI's API. Examples include certain models exposed by inference providers such as [Hugging Face Text Generation Inference (TGI)](https://huggingface.co/docs/text-generation-inference/en/index) and [Google Gemini API](openai-compatible-google-gemini-api.md).
Copy file name to clipboardExpand all lines: articles/api-management/sap-api.md
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,5 +1,5 @@
1
1
---
2
-
title: Import an SAP API by Using the Azure Portal | Microsoft Docs
2
+
title: Import an SAP API by Using the Azure portal | Microsoft Docs
3
3
titleSuffix:
4
4
description: Learn how to import OData metadata from SAP as an API to Azure API Management, either directly or by converting the metadata to an OpenAPI specification.
0 commit comments