Skip to content

Commit 538eb3d

Browse files
committed
[APIM][UUF] Semantic caching note
1 parent 77b0859 commit 538eb3d

4 files changed

Lines changed: 10 additions & 4 deletions

articles/api-management/azure-openai-semantic-cache-lookup-policy.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.collection: ce-skilling-ai-copilot
99
ms.custom:
1010
- build-2024
1111
ms.topic: reference
12-
ms.date: 10/27/2025
12+
ms.date: 02/23/2026
1313
ms.update-cycle: 180-days
1414
ms.author: danlep
1515
---
@@ -23,6 +23,8 @@ Use the `azure-openai-semantic-cache-lookup` policy to perform cache lookup of r
2323
> [!NOTE]
2424
> * This policy must have a corresponding [Cache responses to Azure OpenAI API requests](azure-openai-semantic-cache-store-policy.md) policy.
2525
> * For prerequisites and steps to enable semantic caching, see [Enable semantic caching for LLM APIs in Azure API Management](azure-openai-enable-semantic-caching.md).
26+
> * Because semantic caching returns responses based on similarity (not exact match), it can surface responses that are incorrect, outdated, or unsafe for the current request. Evaluate this feature carefully for your workload and include safeguards.
27+
2628

2729
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2830

articles/api-management/azure-openai-semantic-cache-store-policy.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.collection: ce-skilling-ai-copilot
99
ms.custom:
1010
- build-2024
1111
ms.topic: reference
12-
ms.date: 12/13/2024
12+
ms.date: 02/23/2026
1313
ms.update-cycle: 180-days
1414
ms.author: danlep
1515
---
@@ -23,6 +23,8 @@ The `azure-openai-semantic-cache-store` policy caches responses to Azure OpenAI
2323
> [!NOTE]
2424
> * This policy must have a corresponding [Get cached responses to Azure OpenAI API requests](azure-openai-semantic-cache-lookup-policy.md) policy.
2525
> * For prerequisites and steps to enable semantic caching, see [Enable semantic caching for Azure OpenAI APIs in Azure API Management](azure-openai-enable-semantic-caching.md).
26+
> * Because semantic caching returns responses based on similarity (not exact match), it can surface responses that are incorrect, outdated, or unsafe for the current request. Evaluate this feature carefully for your workload and include safeguards.
27+
2628

2729
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2830

articles/api-management/llm-semantic-cache-lookup-policy.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ ms.collection: ce-skilling-ai-copilot
99
ms.custom:
1010
- build-2024
1111
ms.topic: reference
12-
ms.date: 10/27/2025
12+
ms.date: 02/23/2026
1313
ms.update-cycle: 180-days
1414
ms.author: danlep
1515
---
@@ -23,6 +23,7 @@ Use the `llm-semantic-cache-lookup` policy to perform cache lookup of responses
2323
> [!NOTE]
2424
> * This policy must have a corresponding [Cache responses to large language model API requests](llm-semantic-cache-store-policy.md) policy.
2525
> * For prerequisites and steps to enable semantic caching, see [Enable semantic caching for LLM APIs in Azure API Management](azure-openai-enable-semantic-caching.md).
26+
> * Because semantic caching returns responses based on similarity (not exact match), it can surface responses that are incorrect, outdated, or unsafe for the current request. Evaluate this feature carefully for your workload and include safeguards.
2627
2728
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2829

articles/api-management/llm-semantic-cache-store-policy.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ ms.service: azure-api-management
88
ms.collection: ce-skilling-ai-copilot
99
ms.custom:
1010
ms.topic: reference
11-
ms.date: 12/13/2024
11+
ms.date: 02/23/2026
1212
ms.update-cycle: 180-days
1313
ms.author: danlep
1414
---
@@ -22,6 +22,7 @@ The `llm-semantic-cache-store` policy caches responses to chat completion API re
2222
> [!NOTE]
2323
> * This policy must have a corresponding [Get cached responses to large language model API requests](llm-semantic-cache-lookup-policy.md) policy.
2424
> * For prerequisites and steps to enable semantic caching, see [Enable semantic caching for Azure OpenAI APIs in Azure API Management](azure-openai-enable-semantic-caching.md).
25+
> * Because semantic caching returns responses based on similarity (not exact match), it can surface responses that are incorrect, outdated, or unsafe for the current request. Evaluate this feature carefully for your workload and include safeguards.
2526
2627
[!INCLUDE [api-management-policy-generic-alert](../../includes/api-management-policy-generic-alert.md)]
2728

0 commit comments

Comments
 (0)