← Home

Whitepaper

AI Identity & Access Management Reference Architecture

A layered authorization model for governing users, agents, tools, data, models, and outputs through continuous runtime enforcement.

Core Thesis

Traditional IAM was designed for users, applications, and resources. AI systems introduce agents, models, tools, pipelines, and agent-to-agent workflows that require authorization across the full execution chain.

1. Problem Statement

AI systems do not simply return information. They can retrieve documents, call tools, invoke models, trigger workflows, and generate outputs that influence business decisions. This creates a new identity and authorization problem: the system must understand not only who the user is, but which agent is acting, what capability is being used, what data is involved, why the action is happening, and whether the runtime context is acceptable. Traditional IAM is necessary, but insufficient on its own. AI requires authorization to be evaluated continuously across the execution chain, not only at login or initial application access.

2. Reference Architecture

The AI IAM reference architecture treats agents as first-class identities and evaluates authorization continuously across the chain:

User

Agent

Tool

Data

Model

Output

Each transition is both an execution step and an authorization boundary. The user may be authorized to use an agent, but the agent still needs separate authorization to use a tool. The tool may retrieve data, but the retrieved data must still be filtered by the user’s own entitlements and the agent’s permitted scope. The model may receive context only after data authorization has been enforced. The output must then be checked before release. This creates a complete identity-aware chain from request to response.

3. Layered Authorization Model

The layered model separates authorization into four distinct questions. First, can this user invoke this agent? Second, can this agent use this capability? Third, can this user access the data that may be retrieved or acted upon? Fourth, is this runtime action allowed when user, agent, action, resource, purpose, context, and risk are evaluated together? This separation prevents hidden privilege escalation, where an agent becomes a path around normal access controls. It also allows policy to adapt dynamically as context changes, such as device trust, sensitivity, risk score, or intended purpose.

Layer 1: User → Agent Access

Evaluates authorization dynamically across identity, capability, data, purpose, context, and risk.

Layer 2: Agent → Capability Access

Evaluates authorization dynamically across identity, capability, data, purpose, context, and risk.

Layer 3: User → Data Overlay

Evaluates authorization dynamically across identity, capability, data, purpose, context, and risk.

Layer 4: Runtime Intersection

Evaluates authorization dynamically across identity, capability, data, purpose, context, and risk.

Decision Function

Decision = f(user, agent, capability, action, resource, purpose, context, risk)

4. Secure RAG Enforcement

Retrieval-Augmented Generation introduces a critical control boundary. Data should not enter the model context merely because it is semantically relevant. It must also be authorized. Secure RAG requires policy checks before retrieval, metadata filtering during retrieval, and post-retrieval validation before context injection. This means documents can be filtered by classification, owner, tenant, department, purpose, and user entitlement before the model ever sees them. The goal is simple: prevent unauthorized data from becoming model context. In AI systems, retrieval is not just search. Retrieval is an authorization boundary.

The model never sees data unless both the user and the agent are authorized for the specific purpose, context, and data classification.

5. Policy Model

The policy model must evaluate more than user identity. It must include agent identity, capability scope, action, resource, purpose, runtime context, and risk signals. A user may be allowed to invoke an agent for one purpose but not another. An agent may be allowed to summarize data but not export it. A request may be allowed under normal conditions but require step-up authentication or human review when the resource is sensitive or the session risk is elevated. Policy-as-code makes these decisions explicit, testable, auditable, and reusable across agents and systems.

permit(
  principal == User::"adam",
  action == Action::"InvokeAgent",
  resource == Agent::"finance-agent"
)
when {
  context.purpose == "reporting" &&
  context.risk != "high"
};

In an AWS implementation, Cedar and Amazon Verified Permissions can serve as the business authorization layer, while IAM enforces infrastructure permissions.

6. Observability & Audit

AI IAM must preserve the full identity chain. Every request should produce evidence showing who initiated the action, which agent acted, what delegation or policy context was used, what tools were called, what data was retrieved, what model was invoked, and how the output was controlled. This is essential for governance, incident response, compliance, and trust. If an AI system produces a harmful or unauthorized result, security teams need to reconstruct the execution path. Auditability turns AI execution from an opaque event into a traceable control system.

Identity Context

User, agent, delegation chain, and session attributes.

Policy Decisions

Allow, deny, step-up, evaluated attributes, and decision reason.

Execution Evidence

Retrieved documents, tool calls, model invocation metadata, and output filtering.

7. AWS Implementation

A practical AWS implementation can map the architecture to managed services. Amazon Cognito or IAM Identity Center can authenticate users. API Gateway and Lambda can act as policy enforcement points. Amazon Verified Permissions and Cedar can evaluate business authorization. DynamoDB can store agent registry entries, capability definitions, delegation records, and purpose metadata. S3, Bedrock Knowledge Bases, OpenSearch, or Lake Formation can support controlled retrieval. Amazon Bedrock can provide model invocation and guardrails. CloudWatch, CloudTrail, Security Lake, GuardDuty, and Macie can provide observability, audit, and detection signals.

8. Conclusion

AI IAM extends identity and access management from static access control to continuous authorization across autonomous and semi-autonomous execution chains. It recognizes that agents are not just application features; they are actors that need identity, scope, policy, and accountability. The result is a system where every AI action is identity-aware, purpose-bound, policy-evaluated, risk-adjusted, and auditable. This reference architecture provides a foundation for building AI systems that are not only useful, but governable, explainable, and safe to operate in enterprise environments.