Forum Discussion

armelkamgangfotso's avatar
armelkamgangfotso
Copper Contributor
Jul 07, 2025

Using Microsoft Purview to Identify and Label Sensitive Data Exposed to Generative AI Tools

Hello everyone,
I'm currently working on a data governance initiative and would like to leverage Microsoft Purview to automatically identify and label sensitive data that could potentially be exposed through generative AI (GenAI) tools like Microsoft Copilot, Azure OpenAI services, or other integrated conversational agents.

My main goals are to:

Detect and label sensitive data that may be surfaced or referenced in prompts or AI-generated outputs
Apply sensitivity labels and DLP policies to restrict inappropriate data exposure
Integrate this detection into a broader data loss prevention (DLP) strategy

Here are my questions:

What are the recommended steps to configure Microsoft Purview for monitoring and labeling sensitive data in a GenAI environment?
Is there a way to audit or trace sensitive data usage within interactions involving Copilot or other AI tools?
Do you have any best practices or examples of configuring Sensitive Information Types (SITs) or DLP policies tailored for GenAI scenarios?
Does Microsoft offer native integration between Purview and AI activity, or would we need custom connectors/logs to monitor data exposure?

Any experience, guidance, or references would be greatly appreciated. Thanks in advance for your support!

2 Replies

Resources