|
| 1 | +## Overview |
| 2 | + |
1 | 3 | Audit trails are essential for maintaining trust, compliance, and operational integrity across AIdriven systems. Solution architects must design mechanisms that provide full visibility into _who_ made changes, _what_ changes occurred, and _when_ modifications happened across models, data sources, prompts, configuration objects, and agent workflows. |
2 | 4 |
|
3 | 5 | This unit provides patterns for building robust auditing capabilities across Azure AI Foundry, model lifecycles, and data governance processes. |
4 | 6 |
|
5 | 7 | ## 1. Purpose of audit trails in AI systems |
6 | 8 |
|
7 | | -Audit trails help organizations: |
| 9 | +### Audit trails help organizations: |
8 | 10 |
|
9 | | -Track changes to model versions, parameters, training datasets, and evaluation results. |
| 11 | +- Track changes to model versions, parameters, training datasets, and evaluation results. |
10 | 12 |
|
11 | | -Diagnose issues, anomalies, and regressions. |
| 13 | +- Diagnose issues, anomalies, and regressions. |
12 | 14 |
|
13 | | -Meet regulatory expectations around transparency and accountability. |
| 15 | +- Meet regulatory expectations around transparency and accountability. |
14 | 16 |
|
15 | | -Support investigations during incidents and security reviews. |
| 17 | +- Support investigations during incidents and security reviews. |
16 | 18 |
|
17 | | -Strengthen operational governance by enabling approvals, rollbacks, and lifecycle controls. |
| 19 | +- Strengthen operational governance by enabling approvals, rollbacks, and lifecycle controls. |
18 | 20 |
|
19 | | -A welldesigned audit system is continuous, immutable, and integrated into the AI platform's control plane. |
| 21 | +- A welldesigned audit system is continuous, immutable, and integrated into the AI platform's control plane. |
20 | 22 |
|
21 | 23 | ## 2. Audit trail requirements for model changes |
22 | 24 |
|
23 | 25 | ### Model lifecycle events that must be audited |
24 | 26 |
|
25 | | -Registration of new model versions |
| 27 | +- Registration of new model versions |
26 | 28 |
|
27 | | -Tuning operations such as finetuning, parameter updates, or evaluation runs |
| 29 | +- Tuning operations such as finetuning, parameter updates, or evaluation runs |
28 | 30 |
|
29 | | -Promotion of models between dev/test/prod environments |
| 31 | +- Promotion of models between dev/test/prod environments |
30 | 32 |
|
31 | | -Rollbacks to previous versions |
| 33 | +- Rollbacks to previous versions |
32 | 34 |
|
33 | | -Deployment actions, endpoint configuration, scaling changes |
| 35 | +- Deployment actions, endpoint configuration, scaling changes |
34 | 36 |
|
35 | | -Access attempts to model assets, including unauthorized requests |
| 37 | +- Access attempts to model assets, including unauthorized requests |
36 | 38 |
|
37 | 39 | ### Recommended architectural attributes |
38 | 40 |
|
39 | | -Immutable logs |
| 41 | +- Immutable logs |
40 | 42 |
|
41 | | -Timestamped change records |
| 43 | +- Timestamped change records |
42 | 44 |
|
43 | | -Rolebased attribution (linked to identity provider) |
| 45 | +- Rolebased attribution (linked to identity provider) |
44 | 46 |
|
45 | | -JSONstructured logs for machine parsing |
| 47 | +- JSONstructured logs for machine parsing |
46 | 48 |
|
47 | | -Separation of duties and approval logging |
| 49 | +- Separation of duties and approval logging |
48 | 50 |
|
49 | 51 | ## 3. Audit trail requirements for data changes |
50 | 52 |
|
51 | | -Grounding data, training data, memory stores, semantic indexes, and application logs are core inputs into generative systems. You must audit: |
| 53 | +Grounding data, training data, memory stores, semantic indexes, and application logs are core inputs into generative systems. |
| 54 | + |
| 55 | +### You must audit: |
52 | 56 |
|
53 | | -Data ingestion events |
| 57 | +- Data ingestion events |
54 | 58 |
|
55 | | -Schema changes |
| 59 | +- Schema changes |
56 | 60 |
|
57 | | -Data refresh cycles |
| 61 | +- Data refresh cycles |
58 | 62 |
|
59 | | -Sensitivity label changes |
| 63 | +- Sensitivity label changes |
60 | 64 |
|
61 | | -Sanitization or filtering applied |
| 65 | +- Sanitization or filtering applied |
62 | 66 |
|
63 | | -Data movement across regions |
| 67 | +- Data movement across regions |
64 | 68 |
|
65 | | -Access attempts and approvals |
| 69 | +- Access attempts and approvals |
66 | 70 |
|
67 | | -Architects must ensure logs capture _metadata_, not _content_, to avoid unnecessary exposure of sensitive information. |
| 71 | +- Architects must ensure logs capture _metadata_, not _content_, to avoid unnecessary exposure of sensitive information. |
68 | 72 |
|
69 | 73 | ## 4. Azure AI Foundry controlplane auditing |
70 | 74 |
|
71 | | -Azure AI Foundry provides a centralized control plane for model registration, environment configuration, agent deployment, and diagnostic logging.<br>Key audit features include: |
| 75 | +Azure AI Foundry provides a centralized control plane for model registration, environment configuration, agent deployment, and diagnostic logging. |
| 76 | + |
| 77 | +### Key audit features include: |
72 | 78 |
|
73 | | -### Foundry Activity Logs |
| 79 | +#### Foundry Activity Logs |
74 | 80 |
|
75 | 81 | Track administrative actions across workspaces, registries, and deployments. Logs support export to: |
76 | 82 |
|
77 | | -Azure Monitor |
| 83 | +- Azure Monitor |
78 | 84 |
|
79 | | -Log Analytics |
| 85 | +- Log Analytics |
80 | 86 |
|
81 | | -SIEM tools (such as Sentinel) |
| 87 | +- SIEM tools (such as Sentinel) |
82 | 88 |
|
83 | | -### Foundry Diagnostics and Tracing |
| 89 | +#### Foundry Diagnostics and Tracing |
84 | 90 |
|
85 | 91 | Diagnostics provide traceability of execution across: |
86 | 92 |
|
87 | | -Model calls |
88 | | - |
89 | | -Pipeline activities |
90 | | - |
91 | | -Tool interactions |
92 | | - |
93 | | -Failures, latency spikes, or unexpected behavior |
94 | | - |
95 | | -### Professional Visual (Textbased Diagram) |
96 | | - |
97 | | -+-----------------------------+ |
98 | | - |
99 | | -| Foundry Control Plane | |
100 | | - |
101 | | -+-----------------------------+ |
102 | | - |
103 | | - | Activity Logs |
104 | | - |
105 | | - | Diagnostics |
106 | | - |
107 | | - v |
108 | | - |
109 | | -+-----------------------------+ |
110 | | - |
111 | | -| Azure Monitor / Sentinel | |
112 | | - |
113 | | -+-----------------------------+ |
114 | | - |
115 | | - | Insights | Alerts |
116 | | - |
117 | | - v |
| 93 | +- Model calls |
118 | 94 |
|
119 | | -+-----------------------------+ |
| 95 | +- Pipeline activities |
120 | 96 |
|
121 | | -| Audit Dashboards & Review | |
| 97 | +- Tool interactions |
122 | 98 |
|
123 | | -+-----------------------------+ |
| 99 | +- Failures, latency spikes, or unexpected behavior |
124 | 100 |
|
125 | 101 | ## 5. Designing audit pipelines with tracing |
126 | 102 |
|
127 | 103 | Tracing allows architects to follow execution paths and debug generative AI behaviors. When integrated into audit trails, tracing provides: |
128 | 104 |
|
129 | | -Endtoend visibility of model inference |
| 105 | +- Endtoend visibility of model inference |
130 | 106 |
|
131 | | -Identification of performance bottlenecks |
| 107 | +- Identification of performance bottlenecks |
132 | 108 |
|
133 | | -Attribution of requests to model versions |
| 109 | +- Attribution of requests to model versions |
134 | 110 |
|
135 | | -Correlation of user actions to model activity |
| 111 | +- Correlation of user actions to model activity |
136 | 112 |
|
137 | | -Detection of unusual patterns (loops, excessive token spikes, cascading failures) |
| 113 | +- Detection of unusual patterns (loops, excessive token spikes, cascading failures) |
138 | 114 |
|
139 | 115 | ### Recommended tracing fields |
140 | 116 |
|
141 | | -Correlation ID |
| 117 | +- Correlation ID |
142 | 118 |
|
143 | | -Model version |
| 119 | +- Model version |
144 | 120 |
|
145 | | -Input metadata (nonsensitive) |
| 121 | +- Input metadata (nonsensitive) |
146 | 122 |
|
147 | | -Prompt category or classification |
| 123 | +- Prompt category or classification |
148 | 124 |
|
149 | | -Response latency |
| 125 | +- Response latency |
150 | 126 |
|
151 | | -Downstream tool usage logs |
| 127 | +- Downstream tool usage logs |
152 | 128 |
|
153 | | -Safety evaluation outcomes |
| 129 | +- Safety evaluation outcomes |
154 | 130 |
|
155 | 131 | ## 6. Designing auditready processes |
156 | 132 |
|
157 | 133 | ### Governance workflows to include |
158 | 134 |
|
159 | | -**Approval workflows** for promoting new model versions |
| 135 | +- **Approval workflows** for promoting new model versions |
160 | 136 |
|
161 | | -**Mandatory logging** for all model tuning operations |
| 137 | +- **Mandatory logging** for all model tuning operations |
162 | 138 |
|
163 | | -**Periodic audit reviews** for drift detection |
| 139 | +- **Periodic audit reviews** for drift detection |
164 | 140 |
|
165 | | -**Automated evidence collection** for compliance reporting |
| 141 | +- **Automated evidence collection** for compliance reporting |
166 | 142 |
|
167 | | -**Immutable storage** of audit logs |
| 143 | +- **Immutable storage** of audit logs |
168 | 144 |
|
169 | 145 | ### Retention policies |
170 | 146 |
|
171 | 147 | Define retention requirements with Legal, Compliance, and Information Security teams.<br>Common patterns: |
172 | 148 |
|
173 | | -90 days for lowrisk workloads |
| 149 | +- 90 days for lowrisk workloads |
174 | 150 |
|
175 | | -12-24 months for regulated workloads |
| 151 | +- 12-24 months for regulated workloads |
176 | 152 |
|
177 | | -Indefinite retention for incidentrelated archives |
| 153 | +- Indefinite retention for incidentrelated archives |
178 | 154 |
|
179 | 155 | ## 7. Professional Visual — Audit Trail Architecture |
180 | 156 |
|
@@ -252,8 +228,8 @@ Indefinite retention for incidentrelated archives |
252 | 228 |
|
253 | 229 | ## References |
254 | 230 |
|
255 | | -[https://learn.microsoft.com/en-us/azure/ai-foundry/control-plane/overview?view=foundry](/azure/ai-foundry/control-plane/overview) |
| 231 | +- [https://learn.microsoft.com/en-us/azure/ai-foundry/control-plane/overview?view=foundry](/azure/ai-foundry/control-plane/overview) |
256 | 232 |
|
257 | | -[https://learn.microsoft.com/en-us/training/modules/tracing-generative-ai-app/](/training/modules/tracing-generative-ai-app/) |
| 233 | +- [https://learn.microsoft.com/en-us/training/modules/tracing-generative-ai-app/](/training/modules/tracing-generative-ai-app/) |
258 | 234 |
|
259 | | -[https://learn.microsoft.com/en-us/training/modules/azure-ai-foundry-secure-environment/enable-foundry-diagnostics](/training/modules/azure-ai-foundry-secure-environment/enable-foundry-diagnostics) |
| 235 | +- [https://learn.microsoft.com/en-us/training/modules/azure-ai-foundry-secure-environment/enable-foundry-diagnostics](/training/modules/azure-ai-foundry-secure-environment/enable-foundry-diagnostics) |
0 commit comments