Skip to content

Commit 2611e4a

Browse files
Merge pull request #311163 from MicrosoftDocs/main
Auto Publish – main to live - 2026-01-30 23:00 UTC
2 parents 60904cb + f4ebd2a commit 2611e4a

23 files changed

Lines changed: 965 additions & 425 deletions

articles/api-management/breaking-changes/trusted-service-connectivity-retirement-march-2026.md

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -121,6 +121,7 @@ To do so, set a custom property `Microsoft.WindowsAzure.ApiManagement.Gateway.Ma
121121
"location": "string",
122122
"properties": {
123123
"customProperties": {
124+
// Existing custom properties defined on the service
124125
"Microsoft.WindowsAzure.ApiManagement.Gateway.ManagedIdentity.DisableOverPrivilegedAccess": "True"
125126
}
126127
},
@@ -131,6 +132,9 @@ To do so, set a custom property `Microsoft.WindowsAzure.ApiManagement.Gateway.Ma
131132
}
132133
```
133134

135+
> [!NOTE]
136+
> Existing custom properties, such as ciphers, must be added to the PATCH call as they would otherwise be removed from the service.
137+
134138
The Azure Advisor recommendation should disappear within a day or two of disabling the trusted connectivity on the API Management gateway.
135139

136140
## What is the deadline for the change?

articles/api-management/genai-gateway-capabilities.md

Lines changed: 38 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ author: dlepow
66
ms.service: azure-api-management
77
ms.collection: ce-skilling-ai-copilot
88
ms.topic: concept-article
9-
ms.date: 11/13/2025
9+
ms.date: 01/27/2026
1010
ms.update-cycle: 180-days
1111
ms.author: danlep
1212
ms.custom:
@@ -17,7 +17,7 @@ ms.custom:
1717

1818
[!INCLUDE [api-management-availability-all-tiers](../../includes/api-management-availability-all-tiers.md)]
1919

20-
The *AI gateway* in Azure API Management is a set of capabilities that help you manage your AI backends effectively. These capabilities help you manage, secure, scale, monitor, and govern large language model (LLM) deployments, AI APIs, and Model Context Protocol (MCP) servers that back your intelligent apps and agents.
20+
The *AI gateway* in Azure API Management is a set of capabilities that help you manage your AI backends effectively. Use these capabilities to secure, scale, monitor, and govern AI models, agents, and tools that back your intelligent apps and workloads.
2121

2222
Use the AI gateway to manage a wide range of AI endpoints, including:
2323

@@ -30,7 +30,10 @@ Use the AI gateway to manage a wide range of AI endpoints, including:
3030
:::image type="content" source="media/genai-gateway-capabilities/capabilities-summary.png" alt-text="Diagram summarizing AI gateway capabilities of Azure API Management.":::
3131

3232
> [!NOTE]
33-
> The AI gateway, including [MCP server capabilities](mcp-server-overview.md), extends API Management's existing [API gateway](api-management-key-concepts.md#api-gateway); it isn't a separate offering. Related governance and developer features are in [Azure API Center](../api-center/overview.md).
33+
> The AI gateway, including [MCP server capabilities](mcp-server-overview.md), extends API Management's existing [API gateway](api-management-key-concepts.md#api-gateway); it's not a separate offering. Related governance and developer features are in [Azure API Center](../api-center/overview.md).
34+
35+
> [!NOTE]
36+
> New! AI gateway can now be integrated directly into Microsoft Foundry, enabling you to govern AI models, agents, and tools from within your Foundry environment. Learn more in the [AI gateway in Microsoft Foundry](#ai-gateway-in-microsoft-foundry-preview) section.
3437
3538
## Why use an AI gateway?
3639

@@ -40,18 +43,17 @@ AI adoption in organizations involves several phases:
4043
* Building AI apps and agents that need access to AI models and services
4144
* Operationalizing and deploying AI apps and backends to production
4245

43-
As AI adoption matures, especially in larger enterprises, the AI gateway helps address key challenges, helping to:
46+
As AI adoption matures, especially in larger enterprises, the AI gateway helps address key challenges. It helps you:
4447

4548
* Authenticate and authorize access to AI services
4649
* Load balance across multiple AI endpoints
4750
* Monitor and log AI interactions
4851
* Manage token usage and quotas across multiple applications
4952
* Enable self-service for developer teams
5053

51-
5254
## Traffic mediation and control
5355

54-
With the AI gateway, you can:
56+
By using the AI gateway, you can:
5557

5658
* Quickly import and configure OpenAI-compatible or passthrough LLM endpoints as APIs
5759
* Manage models deployed in Microsoft Foundry or providers such as Amazon Bedrock
@@ -74,13 +76,13 @@ More information:
7476

7577
## Scalability and performance
7678

77-
One of the main resources in generative AI services is *tokens*. Microsoft Foundry and other providers assign quotas for your model deployments as tokens-per-minute (TPM). You distribute these tokens across your model consumers, such as different applications, developer teams, or departments within the company.
79+
One of the main resources in generative AI services is *tokens*. Microsoft Foundry and other providers assign quotas for your model deployments as tokens per minute (TPM). You distribute these tokens across your model consumers, such as different applications, developer teams, or departments within the company.
7880

7981
If you have a single app connecting to an AI service backend, you can manage token consumption with a TPM limit that you set directly on the model deployment. However, when your application portfolio grows, you might have multiple apps calling single or multiple AI service endpoints. These endpoints can be pay-as-you-go or [Provisioned Throughput Units](/azure/ai-services/openai/concepts/provisioned-throughput) (PTU) instances. You need to make sure that one app doesn't use the whole TPM quota and block other apps from accessing the backends they need.
8082

8183
### Token rate limiting and quotas
8284

83-
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of AI service tokens. With this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
85+
Configure a token limit policy on your LLM APIs to manage and enforce limits per API consumer based on the usage of AI service tokens. By using this policy, you can set a TPM limit or a token quota over a specified period, such as hourly, daily, weekly, monthly, or yearly.
8486

8587
:::image type="content" source="media/genai-gateway-capabilities/token-rate-limiting.png" alt-text="Diagram of limiting Azure OpenAI Service tokens in API Management.":::
8688

@@ -122,11 +124,11 @@ More information:
122124
* [Deploy an API Management instance in multiple regions](api-management-howto-deploy-multi-region.md)
123125

124126
> [!NOTE]
125-
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, you should deploy backend AI services in the same regions as your API Management gateways.
127+
> While API Management can scale gateway capacity, you also need to scale and distribute traffic to your AI backends to accommodate increased load (see the [Resiliency](#resiliency) section). For example, to take advantage of geographical distribution of your system in a multiregion configuration, deploy backend AI services in the same regions as your API Management gateways.
126128
127129
## Security and safety
128130

129-
An AI gateway secures and controls access to your AI APIs. With the AI gateway, you can:
131+
An AI gateway secures and controls access to your AI APIs. By using the AI gateway, you can:
130132

131133
* Use managed identities to authenticate to Azure AI services, so you don't need API keys for authentication
132134
* Configure OAuth authorization for AI apps and agents to access APIs or MCP servers by using API Management's credential manager
@@ -139,6 +141,7 @@ More information:
139141
* [Authenticate and authorize access to LLM APIs](api-management-authenticate-authorize-ai-apis.md)
140142
* [About API credentials and credential manager](credentials-overview.md)
141143
* [Enforce content safety checks on LLM requests](llm-content-safety-policy.md)
144+
* [Secure access to MCP servers](secure-mcp-servers.md)
142145

143146

144147
## Resiliency
@@ -163,15 +166,15 @@ More information:
163166

164167
## Observability and governance
165168

166-
API Management provides comprehensive monitoring and analytics capabilities to track token usage patterns, optimize costs, ensure compliance with your AI governance policies, and troubleshoot issues with your AI APIs. Use these capabilities to:
169+
API Management provides comprehensive monitoring and analytics capabilities to track token usage patterns, optimize costs, ensure compliance with your AI governance policies, and troubleshoot problems with your AI APIs. Use these capabilities to:
167170

168-
* Log prompts and completions to Azure Monitor
169-
* Track token metrics per consumer in Application Insights
170-
* View the built-in monitoring dashboard
171-
* Configure policies with custom expressions
172-
* Manage token quotas across applications
171+
* Log prompts and completions to Azure Monitor.
172+
* Track token metrics per consumer in Application Insights.
173+
* View the built-in monitoring dashboard.
174+
* Configure policies with custom expressions.
175+
* Manage token quotas across applications.
173176

174-
For example, you can emit token metrics with the [llm-emit-token-metric](llm-emit-token-metric-policy.md) policy and add custom dimensions you can use to filter the metric in Azure Monitor. The following example emits token metrics with dimensions for client IP address, API ID, and user ID (from a custom header):
177+
For example, you can emit token metrics by using the [llm-emit-token-metric](llm-emit-token-metric-policy.md) policy and add custom dimensions you can use to filter the metric in Azure Monitor. The following example emits token metrics with dimensions for client IP address, API ID, and user ID (from a custom header):
175178

176179
```xml
177180
<llm-emit-token-metric namespace="llm-metrics">
@@ -183,7 +186,6 @@ For example, you can emit token metrics with the [llm-emit-token-metric](llm-emi
183186

184187
:::image type="content" source="media/genai-gateway-capabilities/emit-token-metrics.png" alt-text="Diagram of emitting token metrics using API Management.":::
185188

186-
187189
Also, enable logging for LLM APIs in Azure API Management to track token usage, prompts, and completions for billing and auditing. After you enable logging, you can analyze the logs in Application Insights and use a built-in dashboard in API Management to view token consumption patterns across your AI APIs.
188190

189191
:::image type="content" source="media/api-management-howto-llm-logs/analytics-workbook-small.png" alt-text="Screenshot of analytics for language model APIs in the portal." lightbox="media/api-management-howto-llm-logs/analytics-workbook.png":::
@@ -214,6 +216,23 @@ More information:
214216
* [Azure API Management policy toolkit](https://github.com/Azure/azure-api-management-policy-toolkit/)
215217
* [API Center Copilot Studio connector](../api-center/export-to-copilot-studio.yml)
216218

219+
## AI gateway in Microsoft Foundry (preview)
220+
221+
You can now integrate AI gateway directly into Microsoft Foundry, enabling you to govern AI traffic from within your Foundry environment. When you create or associate an AI gateway instance with your Foundry resource, you can govern, secure, and monitor your Foundry resources through the gateway.
222+
223+
**Models**: Configure token quotas and rate limits directly in the Foundry interface for all model deployments, including Azure OpenAI and other providers.
224+
225+
**Agents**: Register agents running anywhere - Azure, other clouds, or on-premises - into the Foundry control plane for centralized inventory and governance. View telemetry in Foundry or Application Insights, and apply policies such as throttling or content safety.
226+
227+
**Tools**: Register MCP tools hosted across any environment for automatic governance and discovery. Tools appear in the Foundry inventory, ready for consumption by agents.
228+
229+
For advanced scenarios such as custom policies, enterprise networking, or federated gateways, access the full Azure API Management experience while maintaining continuity with Foundry-managed resources.
230+
231+
More information:
232+
233+
* [Enable AI gateway in Microsoft Foundry](/azure/ai-foundry/configuration/enable-ai-api-management-gateway-portal)
234+
* [Register custom agents in Foundry](/azure/ai-foundry/control-plane/register-custom-agent)
235+
* [Govern tools with AI gateway](/azure/ai-foundry/agents/how-to/tools/governance)
217236

218237
## Early access to AI gateway features
219238

@@ -223,7 +242,6 @@ More information:
223242

224243
* [Configure service update settings for your API Management instances](configure-service-update-settings.md)
225244

226-
227245
## Labs and code samples
228246

229247
* [AI gateway capabilities labs](https://github.com/Azure-Samples/ai-gateway)
@@ -240,6 +258,7 @@ More information:
240258

241259
## Related content
242260

261+
* [Blog: AI gateway in Azure API Management is now available in Microsoft Foundry](https://techcommunity.microsoft.com/blog/integrationsonazureblog/ai-gateway-in-azure-api-management-is-now-available-in-microsoft-foundry-preview/4470676)
243262
* [Blog: Introducing AI capabilities in Azure API Management](https://techcommunity.microsoft.com/t5/azure-integration-services-blog/introducing-genai-gateway-capabilities-in-azure-api-management/ba-p/4146525)
244263
* [Blog: Integrating Azure Content Safety with API Management](https://techcommunity.microsoft.com/t5/fasttrack-for-azure/integrating-azure-content-safety-with-api-management-for-azure/ba-p/4202505)
245264
* [Training: Manage your generative AI APIs](/training/modules/api-management)

articles/api-management/secure-mcp-servers.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -80,6 +80,8 @@ For more inbound authorization options and samples, see:
8080

8181
* [MCP server authorization with Protected Resource Metadata (PRM) sample](https://github.com/blackchoey/remote-mcp-apim-oauth-prm)
8282

83+
* [Lab: MCP with protected resource metadata (PRM) authorization](https://github.com/Azure-Samples/AI-Gateway/tree/main/labs/mcp-prm-oauth)
84+
8385
* [Secure Remote MCP Servers using Azure API Management (Experimental)](https://github.com/Azure-Samples/remote-mcp-apim-functions-python)
8486

8587
* [MCP client authorization lab](https://github.com/Azure-Samples/AI-Gateway/tree/main/labs/mcp-client-authorization)

0 commit comments

Comments
 (0)