Skip to content

Commit cc6fc2a

Browse files
authored
Merge pull request #53479 from MicrosoftDocs/NEW-entra-ai-understand
New entra ai understand
2 parents baf4d37 + 083c0f8 commit cc6fc2a

18 files changed

Lines changed: 681 additions & 0 deletions
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.entra-ai-understand.authentication-flows-microsoft-foundry
3+
title: Authentication flows for AI endpoints in Microsoft Foundry
4+
metadata:
5+
title: Authentication flows for AI endpoints in Microsoft Foundry
6+
description: "Authentication flows for AI endpoints in Microsoft Foundry."
7+
ms.date: 2/13/2026
8+
author: wwlpublish
9+
ms.author: riswinto
10+
ms.topic: unit
11+
azureSandbox: false
12+
labModal: false
13+
durationInMinutes: 5
14+
content: |
15+
[!include[](includes/authentication-flows-microsoft-foundry.md)]
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.entra-ai-understand.common-identity-misconfigurations-ai
3+
title: Common identity misconfigurations in AI deployments
4+
metadata:
5+
title: Common identity misconfigurations in AI deployments
6+
description: "Common identity misconfigurations in AI deployments."
7+
ms.date: 2/13/2026
8+
author: wwlpublish
9+
ms.author: riswinto
10+
ms.topic: unit
11+
azureSandbox: false
12+
labModal: false
13+
durationInMinutes: 6
14+
content: |
15+
[!include[](includes/common-identity-misconfigurations-ai.md)]
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.entra-ai-understand.human-workload-identities-ai-environments
3+
title: Human and workload identities in AI workloads
4+
metadata:
5+
title: Human and workload identities in AI workloads
6+
description: "Human and workload identities in AI workloads."
7+
ms.date: 2/13/2026
8+
author: wwlpublish
9+
ms.author: riswinto
10+
ms.topic: unit
11+
azureSandbox: false
12+
labModal: false
13+
durationInMinutes: 5
14+
content: |
15+
[!include[](includes/human-workload-identities-ai-environments.md)]
Lines changed: 15 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,15 @@
1+
### YamlMime:ModuleUnit
2+
uid: learn.wwl.entra-ai-understand.identity-control-layer-ai-solutions
3+
title: Identity as the control layer for AI solutions
4+
metadata:
5+
title: Identity as the control layer for AI solutions
6+
description: "Identity as the control layer for AI solutions."
7+
ms.date: 2/13/2026
8+
author: wwlpublish
9+
ms.author: riswinto
10+
ms.topic: unit
11+
azureSandbox: false
12+
labModal: false
13+
durationInMinutes: 5
14+
content: |
15+
[!include[](includes/identity-control-layer-ai-solutions.md)]
Lines changed: 62 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,62 @@
1+
Understanding that identity governs access isn't enough. You also need to understand how an identity proves who it is when calling an AI endpoint.
2+
3+
Every request to an AI endpoint must include valid authentication credentials. Microsoft Foundry relies on industry-standard OAuth 2.0 flows and Microsoft Entra ID to issue and validate access tokens.
4+
5+
At a high level, authentication answers three questions:
6+
7+
- Who is calling the model?
8+
- From where is the request originating?
9+
- What credential is being presented?
10+
11+
## OAuth 2.0 and token issuance
12+
13+
AI endpoints use OAuth 2.0 to authenticate callers. When a user, application, or workload attempts to access a protected AI resource, Microsoft Entra ID evaluates the request and issues an access token if authentication succeeds.
14+
15+
That token represents the authenticated identity. It contains claims about:
16+
17+
- The identity
18+
- The tenant
19+
- Assigned roles or permissions
20+
- Token lifetime
21+
22+
The caller includes the access token in requests to the AI endpoint. The service validates the token before processing the request.
23+
24+
For example, a web application that calls an AI endpoint uses its service principal to request a token from Microsoft Entra ID. If authentication succeeds, Microsoft Entra issues an access token. The application includes that token in its request, and the AI endpoint validates it before processing the prompt.
25+
26+
If the token is invalid, expired, or improperly scoped, the request fails before reaching the model.
27+
28+
## Interactive and non-interactive authentication
29+
30+
Authentication flows differ depending on who or what is calling the endpoint.
31+
32+
Interactive authentication typically involves a human user. For example:
33+
34+
- A developer signs in to deploy a model through Microsoft Foundry.
35+
- An administrator configures AI resources in the Azure portal.
36+
37+
In these cases, the user authenticates through Microsoft Entra ID, and an access token is issued for that session.
38+
39+
Non-interactive authentication is used by applications and workloads. Examples include:
40+
41+
- A web application calling a model endpoint
42+
- An automation workflow deploying updates
43+
- A backend service retrieving embeddings
44+
45+
In these scenarios, authentication occurs without human interaction. The application uses a service principal or managed identity to obtain an access token from Microsoft Entra ID.
46+
47+
At runtime, the AI endpoint doesn't distinguish between human and application requests. It validates the token and evaluates authorization based on the identity represented in that token.
48+
49+
## Token flow between client and AI endpoint
50+
51+
The authentication sequence follows a consistent pattern:
52+
53+
1. The client requests an access token from Microsoft Entra ID.
54+
1. Microsoft Entra ID authenticates the caller and issues a token if the request is valid.
55+
1. The client includes the token in the request to the AI endpoint.
56+
1. The endpoint validates the token before evaluating authorization.
57+
58+
Only after successful authentication and authorization does the service process the model request.
59+
60+
This flow ensures that every AI interaction in Microsoft Foundry is tied to a verified identity and evaluated against assigned permissions.
61+
62+
Understanding this sequence prepares you to analyze authentication failures, token misuse, and access control misconfigurations in AI environments. It also clarifies how different identity types authenticate and why that distinction matters.
Lines changed: 85 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,85 @@
1+
Strong authentication and well-defined roles don't guarantee secure AI workloads. Identity design decisions determine whether access boundaries hold under pressure.
2+
3+
Identity design failures cause most AI security incidents, not flaws in the model itself. Common drivers include misconfigured identities, excessive permissions, and weak enforcement controls.
4+
5+
Understanding common misconfigurations helps you recognize risk patterns early.
6+
7+
## Overprivileged application identities
8+
9+
Service principals and managed identities often receive broad permissions for convenience.
10+
11+
For example:
12+
13+
- Assigning contributor access at the subscription level
14+
- Granting both management and data plane permissions when only one is required
15+
- Reusing a single application identity across multiple workloads
16+
17+
Because application identities operate without human interaction, excessive permissions can enable automated misuse at scale.
18+
19+
When an overprivileged identity is compromised, the blast radius extends to every resource within its assigned scope.
20+
21+
## Broad scope assignments
22+
23+
Scope selection directly affects blast radius.
24+
25+
Assigning roles at the subscription level might simplify administration, but it broadens access beyond what's necessary. A compromised identity with subscription-level permissions can modify or access multiple AI resources across environments.
26+
27+
Limiting scope to a specific resource group or AI endpoint narrows the scope of access and aligns more closely with least-privilege principles.
28+
29+
Convenience shouldn't override scope discipline.
30+
31+
## Shared credentials and secret management
32+
33+
Application identities that rely on shared secrets introduce additional risk.
34+
35+
Embedding credentials in code, storing them in unsecured configuration files, or failing to rotate secrets increases the likelihood of compromise.
36+
37+
Managed identities reduce some of this risk by removing the need to manage credentials directly. However, they don't eliminate the need for proper role assignment and scope control.
38+
39+
Secret management failures often surface long after initial deployment, when credentials are reused across environments or teams.
40+
41+
## Mixing development and production identities
42+
43+
Using the same identity for development and production workloads expands the potential scope of a compromise.
44+
45+
Development environments often have broader permissions and less restrictive controls. If that identity also has production access, a compromise in development can affect production systems.
46+
47+
Separating identities by environment limits lateral movement and narrows the scope of a compromise.
48+
49+
Environment boundaries should be reflected in identity boundaries.
50+
51+
## Token handling and leakage risks
52+
53+
Access tokens represent authenticated identities. If tokens are logged, exposed in client-side code, or transmitted insecurely, they can be replayed until expiration.
54+
55+
Short token lifetimes and secure storage practices reduce this risk. Monitoring unusual token usage patterns also helps detect misuse.
56+
57+
Authentication doesn't eliminate risk. It defines the boundary of trust.
58+
59+
## Missing Conditional Access enforcement
60+
61+
Authentication alone doesn't guarantee secure access conditions.
62+
63+
If Conditional Access policies aren't applied to relevant identities, attackers might authenticate successfully without meeting device compliance, location, or multifactor requirements.
64+
65+
Conditional Access strengthens identity enforcement by evaluating context in addition to credentials.
66+
67+
Without it, identity security relies solely on credential protection.
68+
69+
Certain operational symptoms often point back to identity design decisions. Unexpected model deployments can indicate excessive management plane permissions, while unusual model invocation patterns or data access spikes might suggest overly broad data plane access. When investigating AI incidents, examining identity scope and role assignments is often more revealing than inspecting the model itself.
70+
71+
## Closing the loop
72+
73+
Identity architecture defines who can access AI resources, what they can do, and where that access applies.
74+
75+
Misconfigurations weaken those boundaries.
76+
77+
Designing identity intentionally means:
78+
79+
- Assigning only required permissions
80+
- Limiting scope appropriately
81+
- Separating environments
82+
- Protecting credentials
83+
- Enforcing contextual access controls
84+
85+
AI security depends on identity discipline. The tools that monitor posture and detect risk build on these foundations, but they don't replace them.
Lines changed: 69 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,69 @@
1+
AI endpoints don't distinguish between human and application requests at runtime. They validate tokens and evaluate permissions based on the identity presented.
2+
3+
For security design, the type of identity matters.
4+
5+
Understanding the differences between **human**, **application**, and **managed identities** helps you assign permissions correctly and reduce risk.
6+
7+
## Human identities
8+
9+
Human identities represent individual users authenticated through Microsoft Entra ID.
10+
11+
These identities are used when a person directly interacts with AI resources. Examples include:
12+
13+
- A developer deploying or updating a model
14+
- An administrator configuring resource settings
15+
- A security analyst reviewing posture or activity
16+
17+
Human identities are typically authenticated through interactive sign-in. Conditional Access, multifactor authentication, and device compliance policies commonly apply to these sessions.
18+
19+
Because human identities represent real individuals, their permissions should reflect job responsibilities and follow least-privilege principles.
20+
21+
Overassigning permissions to user accounts increases risk, particularly in the management plane.
22+
23+
## Application identities
24+
25+
Application identities represent software rather than people. In Microsoft Entra ID, these are typically implemented as service principals associated with application registrations. An application registration defines the application globally, while a service principal represents that application as a security identity within a specific tenant.
26+
27+
They're used when software calls an AI endpoint directly. Examples include:
28+
29+
- A web application submitting prompts
30+
- A backend service retrieving embeddings
31+
- An automation workflow deploying updates
32+
33+
Authentication occurs through non-interactive flows. The application presents credentials, and Microsoft Entra ID issues an access token for that identity.
34+
35+
Application identities shouldn't use shared secrets embedded in code or configuration without proper protection. Excessive permissions assigned to a service principal can allow broad, automated misuse of AI resources.
36+
37+
## Managed identities
38+
39+
Managed identities are a specialized type of application identity. They're designed to reduce credential management overhead.
40+
41+
When you enable a managed identity on an Azure resource, Microsoft Entra ID automatically creates and manages the associated identity. The platform handles credential rotation and lifecycle management.
42+
43+
Managed identities are appropriate when:
44+
45+
- An Azure-hosted resource needs to access an AI endpoint
46+
- You want to avoid storing credentials in application code
47+
- You want tighter integration with Azure role assignments
48+
49+
Managed identities reduce the risk associated with secret management. However, they still require careful role assignment and scope control.
50+
51+
## Choosing the appropriate identity model
52+
53+
The appropriate identity type depends on who or what is accessing the AI resource.
54+
55+
Use a human identity when a person is directly performing administrative or development tasks.
56+
57+
Use an application identity when software must authenticate independently.
58+
59+
Use a managed identity when the workload runs in Azure and can rely on platform-managed credentials.
60+
61+
Each identity type introduces different security considerations:
62+
63+
- Human identities require strong authentication controls.
64+
- Application identities require strict permission boundaries.
65+
- Managed identities require careful scope assignment.
66+
67+
The identity type you choose affects the potential blast radius if that identity is compromised.
68+
69+
Understanding these distinctions allows you to design access intentionally rather than defaulting to broad or convenient configurations. The next step is defining what those identities are actually permitted to do and where those permissions apply.
Lines changed: 48 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,48 @@
1+
Securing AI workloads starts with identity. Before configuring roles or conditional access policies, you need to understand how authentication and authorization govern every interaction with an AI service.
2+
3+
Every prompt submitted to a model, every deployment, every configuration change, and every API call requires an authenticated identity. AI services are designed to enforce identity-based access and don't support anonymous interaction by design.
4+
5+
If identity controls access to Azure resources, it also determines how AI systems are used. That makes identity the control layer for AI security.
6+
7+
Before evaluating model safety settings or data protections, start with a more fundamental question:
8+
9+
**Who is allowed to access the AI service, and under what conditions?**
10+
11+
## Why identity comes first in AI security
12+
13+
AI services expose authenticated endpoints. Access to AI services isn't anonymous. A user, application, or managed workload must present credentials that Microsoft Entra ID validates.
14+
15+
Once authenticated, authorization determines what that identity can do:
16+
17+
- Deploy or modify a model
18+
- Invoke a model endpoint
19+
- Access training data
20+
- Change configuration settings
21+
22+
These decisions occur before the AI service processes a request.
23+
24+
Authentication verifies identity. Authorization enforces assigned permissions. Only after both steps succeed does the AI service process the request.
25+
26+
In many real-world deployments, identity design issues cause more security failures than the model itself. Common examples include:
27+
28+
- Excessive permissions
29+
- Broad role assignments
30+
- Shared credentials
31+
- Unmanaged service principals
32+
33+
When identity is misconfigured, the AI environment inherits that weakness.
34+
35+
## Identity across AI development and runtime
36+
37+
AI solutions are often developed and deployed using services like Microsoft Foundry. These services integrate directly with Microsoft Entra ID for authentication and authorization.
38+
39+
Identity is present at every stage:
40+
41+
- Developers authenticate to create and configure resources.
42+
- Applications authenticate to call model endpoints.
43+
- Automation workflows authenticate to deploy updates.
44+
- Security teams authenticate to review posture and investigate activity.
45+
46+
Security platforms like Microsoft Defender for Cloud rely on identity context to evaluate risk and surface misconfigurations. Without a clear identity architecture, posture insights lack meaningful context.
47+
48+
Secure AI access rests on four elements: identity, authentication, authorization, and scope. Each layer builds on the previous one. When identity design is sound, these layers work together to create predictable and enforceable access boundaries.
Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,25 @@
1+
Most AI security conversations focus on model behavior, data exposure, or runtime misuse. In AI workloads running in Azure, identity architecture often determines whether those risks are even possible.
2+
3+
AI workloads don't change how Microsoft Entra ID works. What changes is how identity decisions play out once models are deployed, invoked at scale, and integrated into automated pipelines. The separation between management and data plane operations, combined with scoped role assignments, introduces risk patterns that aren't obvious in traditional application deployments.
4+
5+
In AI environments, identity isn't just about signing in to a portal. It defines what actions are possible across deployment and runtime:
6+
7+
- Who can deploy or modify models
8+
- Who can invoke endpoints and retrieve data
9+
- How applications and services authenticate during execution
10+
- How far permissions extend across subscriptions, resource groups, and individual AI resources
11+
12+
Small gaps between identity type, role assignment, and scope can quietly expand the blast radius. Those gaps often surface only during an incident or investigation.
13+
14+
## Learning objectives
15+
16+
By the end of this module, you'll be able to:
17+
18+
- Explain identity as the control layer for AI solutions in Azure
19+
- Distinguish between management plane and data plane access in AI workloads
20+
- Describe authentication flows used by AI endpoints integrated with Microsoft Entra ID
21+
- Distinguish between human and workload identities
22+
- Interpret role assignments and scope boundaries across AI resources
23+
- Recognize common identity design patterns that increase AI risk
24+
25+
Once identity behavior in AI workloads is clear, access decisions become intentional rather than automatic. That clarity is essential when designing role assignments, enforcing least privilege, and applying Conditional Access in AI environments.

0 commit comments

Comments
 (0)