|
| 1 | +Once you decide how to start a policy, the next step is defining what the policy should detect. Detection choices shape how accurate a policy is, how often it triggers, and how much noise it generates over time. |
| 2 | + |
| 3 | +Good detection isn't about catching everything. It's about identifying the situations that matter most and responding consistently when they occur. |
| 4 | + |
| 5 | +## Ensure detection matches the risk |
| 6 | + |
| 7 | +In Microsoft Purview, detection isn't based on a single signal. Policies evaluate different kinds of signals that describe what the data is, how it's classified, and how it's being used. |
| 8 | + |
| 9 | +These signals generally fall into three categories: |
| 10 | + |
| 11 | +- **Content-based signals** focus on what the data contains, like identifying specific patterns or content types. |
| 12 | + |
| 13 | + These signals are useful when detection needs to be driven by the data itself rather than prior classification. They work well when patterns, keywords, or content structure matter more than how the data was labeled. |
| 14 | + |
| 15 | +- **Classification-based signals** rely on prior classification decisions, like labels applied by users or automation. |
| 16 | + |
| 17 | + These signals are useful when protection should follow the data consistently, regardless of where it appears or how it's used. |
| 18 | + |
| 19 | +- **Context-based signals** describe how data is handled, like the action being taken or where the data is sent. |
| 20 | + |
| 21 | + These signals focus on actions, destinations, and user context, and they help narrow enforcement to situations that actually represent risk. |
| 22 | + |
| 23 | +Effective policies often combine signals from more than one category. Choosing the right mix depends on how reliably data is classified and what behavior the policy is meant to address. |
| 24 | + |
| 25 | +## Combine conditions to improve accuracy |
| 26 | + |
| 27 | +Single detection conditions can work for simple scenarios, but they often lack context. Combining conditions helps narrow enforcement to situations that actually represent risk. |
| 28 | + |
| 29 | +For example, combining content-based detection with: |
| 30 | + |
| 31 | +- A specific action |
| 32 | +- A destination or location |
| 33 | +- A user or group scope |
| 34 | + |
| 35 | +Can significantly reduce unnecessary triggers. |
| 36 | + |
| 37 | +The goal isn't complexity for its own sake. It's clarity. Each condition should contribute meaningfully to the scenario you're trying to address. |
| 38 | + |
| 39 | +## Balance coverage with precision |
| 40 | + |
| 41 | +Broad detection increases coverage, but it also increases the chance of false positives. Narrow detection improves precision, but it can miss edge cases. |
| 42 | + |
| 43 | +Early in policy creation, it's often better to favor clarity over completeness. A policy that triggers reliably in fewer scenarios is easier to validate and refine than one that fires constantly with mixed results. |
| 44 | + |
| 45 | +Detection can always be expanded later. Noise is harder to undo. |
| 46 | + |
| 47 | +## Define what "good enough" detection looks like |
| 48 | + |
| 49 | +Detection doesn't have to be perfect on day one. What matters is whether it reliably identifies the behavior you care about without disrupting normal work. |
| 50 | + |
| 51 | +"Good enough" detection usually means: |
| 52 | + |
| 53 | +- The policy triggers when expected |
| 54 | +- Results are understandable |
| 55 | +- False positives are limited and explainable |
| 56 | + |
| 57 | +This creates a strong foundation for validation and tuning. |
| 58 | + |
| 59 | +## Account for how detection choices affect false positives |
| 60 | + |
| 61 | +Detection decisions made early often determine where false positives appear later. Overly broad conditions, weak context, or reliance on unreliable signals can all contribute to unnecessary enforcement. |
| 62 | + |
| 63 | +Being intentional about detection upfront reduces the need for heavy tuning after deployment. |
| 64 | + |
| 65 | +## Consider scenarios where data is reused or transformed |
| 66 | + |
| 67 | +Some scenarios are more complex than simple sharing or copying. When sensitive data is reused or transformed, detection becomes more important. |
| 68 | + |
| 69 | +This includes workflows where: |
| 70 | + |
| 71 | +- Content is rewritten or summarized |
| 72 | +- Data is combined with other inputs |
| 73 | +- Sensitive information appears in generated responses |
| 74 | + |
| 75 | +In these cases, detection quality matters more than aggressive enforcement. Clear, accurate detection helps ensure policies respond to real risk instead of incidental use. |
| 76 | + |
| 77 | +With detection defined, the next step is deciding where the policy should apply and who it should affect. |
0 commit comments