๐Ÿ“Œ Author's note: This site synthesises the author's own understanding from publicly available Microsoft documentation, official Microsoft Security blog posts, RSAC 2026 announcements, and insights from Microsoft Security professionals and MVPs. It is independent and not affiliated with or endorsed by Microsoft. Microsoft updates products and documentation frequently โ€” always verify current status directly with Microsoft before making architecture or purchasing decisions.
UPDATED ยท FIELD RESEARCH ยท MARCH 2026

How AI Breaks
Traditional Security

Traditional security was built for users, endpoints, and applications. AI agents violate all three assumptions. Field research from Microsoft Security professionals and Microsoft's own agent misconfiguration research reveals the real-world risks are worse than most organisations realise.

Agent Properties & Risk

Properties That Create New Attack Surface

PropertyCapability UpsideSecurity DownsideRisk Severity
Self-initiatingAutomates workflows without human promptsMay take unintended actions outside guardrailsHIGH
PersistentContinuous value; handles tasks 24/7Over-permissioning drift; undetected misuse; orphaned agentsHIGH
OpaqueAbstracts complexity; simplifies workflowsLLM black-box; hard to audit; LLM non-determinism makes output unpredictableHIGH
ProlificLow-code / no-code creation accelerates adoptionShadow agents; sprawl; most existing Copilot Studio agents are Classic โ€” outside Entra security perimeter entirelyCRITICAL
Tool-invokingReal actions: email, APIs, file writePrompt injection converts to real-world harm; MCP tools extend this to any connected systemCRITICAL
Context-consumingRich reasoning over enterprise dataSensitive data enters AI context โ€” new exfiltration surfaceCRITICAL
Maker-authenticatedCreator can configure deep integration at build timeCopilot Studio agents authenticate as their maker, not the user โ€” maker's full permissions extended to every user who interacts with the agentCRITICAL
AutonomousActs independently without per-action human approvalOverprivileged, manipulated, or misaligned agents can act as "double agents" โ€” working against the outcomes they were built to support (Microsoft ZT4AI framing). Agents with excessive permissions and no guardrails are the primary risk vector.CRITICAL
โš  The Maker Credentials Problem โ€” Worse Than OBO

Our Identity page covers the OBO (On-Behalf-Of) token problem. Copilot Studio introduces a more dangerous variant: maker credentials. The agent authenticates to connected services as the person who built it โ€” not the person using it. If a developer with admin rights builds an agent and shares it org-wide with one toggle, every employee in the organisation can interact with it using the maker's admin-level permissions. This is the most widespread and underappreciated privilege escalation risk in current enterprise AI deployments. Field research by Microsoft Security MVP Derk van der Woude confirms this pattern is common in production environments.

๐Ÿ“Œ Going deeper on Copilot Studio risks

The specific Copilot Studio risk patterns โ€” no-auth agents, org-wide sharing, Classic vs Modern agents, maker credentials, ownerless agents, name sync bug โ€” are covered in detail on the Copilot Studio vs Microsoft Foundry page alongside the five authentication patterns and 30-minute audit KQL. For detection and remediation runbooks, see Playbooks.

๐Ÿ“Š Not all AI security incidents are attacks โ€” Microsoft research, March 2026

Microsoft's Secure Access in the Age of AI report found a near-even split: 53% of AI-related access incidents were malicious, 47% were accidental. Accidental incidents are driven by complexity, unclear ownership, and misaligned controls โ€” not adversaries. As agents are deployed faster than policies can be updated and permissions are configured broadly "to make sure they work", unintentional misuse escalates quickly. The risk taxonomy below covers both categories.

Additionally: 97% of organisations experienced an identity or network access incident in the past year, and 70% of those were tied to AI-related activity. Source: Microsoft Entra Blog, March 2026

Risk Taxonomy

AI-Specific Risk Categories

RiskDescriptionWho Owns ItPrimary Microsoft Control
Agent sprawlNo inventory of deployed agents; no lifecycle ownershipIT / SecurityAgent 365 โš  per-user license
Classic agents โ€” outside Entra perimeterMost existing Copilot Studio agents are Classic Service Principals with no Entra security product coverageIAM / SecurityMigration to Modern Agents โš  tool not yet available
Maker credentialsCopilot Studio agents authenticate as their builder โ€” maker's permissions extended to all users of the agentIAM / AppSecPower Platform Managed Environments; enforce end-user auth per agent
No-auth agentsAgents set to no authentication โ€” accessible to anyone in Teams with no loginIT / SecurityAIAgentsInfo KQL detection; Power Platform admin enforcement
Org-wide sharingOne toggle exposes agent to all employees โ€” compounds with maker credentialsIT / SecurityPower Platform Managed Environments โ€” set sharing limits
Over-permissioned accessAgents granted broad access; OBO inherits user's full rightsIAM / SecurityEntra Agent ID โš  preview, Modern Agents only
Shadow AI / pluginsBusiness users deploy unsanctioned AI tools and MCP servers outside IT oversightIT / CASBDefender for Cloud Apps + Entra Internet Access GA Mar 31 2026
MCP tool misuseAgents invoke real enterprise tools via MCP โ€” now via official Microsoft MCP server catalogAppSec / SecurityFoundry Guardrails โš  preview + Defender for Cloud Apps
AI model supply chainPretrained models from registries (Hugging Face, Azure ML) may carry embedded malware or backdoors. Training data can be poisoned before it reaches the pipeline. Build-time risks that traditional AppSec doesn't cover.AppSec / ML EngAI Model Scanning in Defender for Cloud GA ยท RSAC 2026
Prompt injection / XPIAMalicious inputs hijack agent behaviour mid-taskAppSec / SOCPrompt Shields + Entra Internet Access Prompt Injection Protection GA Mar 31 2026
Data leakageSensitive data enters AI context; exfiltrated via outputs or promptsDLP / CompliancePurview DSPM + Purview DLP for Copilot GA Mar 31 2026
Ownerless agentsNo accountable owner โ€” agents persist indefinitely with no governance reviewIT / IAMPower Platform Inventory; AIAgentsInfo Advanced Hunting
๐Ÿ“Œ Zero Trust for AI

How the three Zero Trust principles โ€” Verify Explicitly, Use Least Privilege, Assume Breach โ€” apply specifically to AI agents, the three-stage maturity model, and the 12 highest-priority ZT Workshop controls are covered on the Frameworks page.

Classifying your estate

Risk tier methodology โ€” apply H / M / L to every agent

Risk categories describe what can go wrong. This tier rubric tells you which agents to remediate first. Apply the criteria below to every agent in your inventory register โ€” the tier drives both the remediation timeline and the depth of subsequent controls (red teaming scope, governance review cadence, sponsor accountability).

TierCriteria (any of)Required actionGovernance cadence
HIGH
Priority 1
โ€ข No authentication
โ€ข Maker credentials at agent or connector level
โ€ข Org-wide sharing
โ€ข No assigned owner
โ€ข Handles regulated data (PII, financial, health, citizen records)
Remediate or block within 14 days. Full red team engagement before production. Sentinel Analytics Rule alerting on any change.Reviewed at every Agent Lifecycle Board (monthly)
MEDIUM
Priority 2
โ€ข Authenticated but broad connector access (SharePoint, Exchange, Teams)
โ€ข Sensitive but not regulated data connectors
โ€ข Named owner present but no business sponsor
โ€ข Shared with large group (50+) but not org-wide
Scope review within 30 days. DLP coverage validated. Annual focused red team on prompt injection & data exfiltration.Quarterly governance sweep
LOW
Monitor
โ€ข Delegated end-user authentication
โ€ข Named users or small group sharing
โ€ข Scoped data access (single team site, single connector)
โ€ข Both owner and sponsor assigned
Document. Include in quarterly inventory check. Regression red team only on significant change (connector, tool, system prompt).Annual audit + change-triggered review
โš  Critical: the tier is the highest match, not the average

An agent that meets one HIGH criterion and four LOW criteria is still HIGH. Risk does not average down. A no-auth agent serving a tiny team and reading only a single SharePoint list is still HIGH because the no-auth condition alone makes it externally reachable. Apply the criteria as a screen, not a score.

๐Ÿ“Œ Where the tier is used

Phase 1 inventory output โ†’ Phase 2 governance sequence (remediate HIGH first) โ†’ Phase 4 red team scope (Tier 1/2/3 maps to High/Medium/Low) โ†’ Phase 6 board reporting (tier distribution + trend). See the six-phase rollout on Strategy for the full sequence.

Trust & Safety

AI Trust and Safety assurance โ€” distinct from security testing

For agents that interact directly with citizens, vulnerable users, or make consequential decisions (benefits, healthcare, financial outcomes, public-facing services), security testing alone is not sufficient. An AI Trust and Safety assessment โ€” typically using a recognised safety assurance methodology such as Adelard's safety case methodology โ€” validates that the agent is trustworthy, reliable, and dependable across its full data pipeline. This is a separate assurance discipline from penetration testing or red teaming.

DisciplineWhat it testsOutput
Security testing / red teamingAdversarial robustness โ€” prompt injection, data exfiltration, tool chain manipulation, jailbreak, model extractionVulnerabilities & exploit paths ยท OWASP LLM Top 10 coverage
AI Trust and Safety assuranceReliability, dependability, fairness, safety-of-use โ€” including failure modes for vulnerable users, edge cases, hallucination tolerance, and traceability of decisionsSafety case ยท auditable assurance argument ยท evidence pack for regulators
Responsible AI evaluationHarms โ€” bias, toxicity, manipulation, discriminatory outputs โ€” typically via Foundry evaluations and Content SafetyHarm taxonomy coverage ยท evaluation telemetry
๐Ÿ“Œ When to commission a Trust & Safety assessment

Required for: agents interacting directly with the public, agents making decisions about benefits or eligibility, agents in healthcare or safeguarding contexts, agents handling vulnerable users. Recommended for: any agent classified as HIGH risk per the tier methodology above, plus any agent newly subject to EU AI Act Annex III high-risk classification. The output is an auditable safety case โ€” not a posture score โ€” suitable for regulator submission alongside the security evidence pack.

STAY UPDATED
Get notified when Microsoft AI security changes
Monthly updates on new controls, GA announcements, and critical gaps โ€” direct to your inbox.
Subscribe to updates โ†’
aiagentsecurity.substack.com ยท Free ยท No spam