Part 6. The Full AI Security Architecture: Secure the Brain, Secure the Hands, Secure the Memory

Kumar Mehta
By Kumar Mehta
Founder and CDO, Versa Networks
April 1, 2026
in
Share
Follow
  • Discovery
  • Control
  • Prompt inspection
  • Model governance
  • Tool governance

This post ties these pieces into one system that a real enterprise can run.

The simplest mental model

AI security becomes easier to understand when you separate “thinking” from “doing.”

  • The model is where AI reasons and produces text. This is the “brain.”
  • Tools are where AI takes action inside your environment. This is the “hands.”

That means you need enforcement in both places. If you only secure model calls but ignore tool calls, your AI can still take unsafe actions. If you only secure tools but ignore model access, you can still leak sensitive data to an unapproved model provider.

What the full stack looks like

A practical reference architecture works like this:

  • Model traffic flows through a Model Gateway. This is where you apply identity-based access, approved model lists, logging, and quotas.
  • Prompts and responses are inspected by an LLM WAF layer. This is where you detect injection attempts, redact sensitive data, and block unsafe requests.
  • Tool traffic flows through an MCP Gateway. This is where you approve tools, restrict high-risk actions, and log every tool call.
  • Tool outputs are treated as untrusted by default, and are inspected before they are placed back into the model context.

As an example, a user asks an AI assistant to “summarize this customer escalation.” The assistant retrieves Slack messages and internal notes. Those tool outputs may contain sensitive details. Before the assistant uses them to generate a response, the system inspects them and redacts anything that should not leave the company.

A real-world incident chain

Here is a common incident chain: 1. A user asks the assistant to summarize an email or document. 2. The document contains hidden malicious instructions. 3. The model attempts to follow those malicious instructions. 4. The agent tries to call tools. 5. Sensitive data is exposed or unsafe actions are taken.

Instead, the secure architecture prevents the attack chain by inspecting prompts for suspicious instruction patterns, restricting tool access through the MCP Gateway unless explicitly authorized, redacting sensitive data before any output leaves the system, and logging every step so security teams can quickly investigate and respond if needed. Here is how the secure architecture stops the chain:

  • Prompt inspection detects suspicious instruction patterns.
  • The MCP Gateway blocks tool access unless explicitly permitted.
  • Sensitive data is redacted before any output leaves the system.
  • Every step is logged so security teams can investigate quickly.

The 90-day rollout plan that works

Phase 1 (first 30 days):

  • Visibility
  • Discover model usage
  • Discover tool usage
  • Establish audit logs and traceability

Phase 2 (days 31–60): Control

  • Enforce allowed models
  • Enforce identity-bound access
  • Enforce basic redaction rules

Phase 3 (days 61–90): Runtime enforcement

  • Enable prompt injection defenses
  • Restrict write tools and require approvals
  • Build incident response playbooks for AI events

What “good” looks like

A mature AI security program provides clear visibility into which models are used and by whom, prevents sensitive data from being sent to unapproved models, inspects prompts and responses for injection attacks and data leakage, governs what tools agents can access and the actions they can take, and ensures all significant AI activity is logged and auditable. The key takeaway is that GenAI security becomes manageable when it’s treated like a real system—by securing the brain (models), the hands (tools), and the memory (data).

Final takeaway

GenAI security becomes manageable when you build it like a real system:

Secure the brain (models). Secure the hands (tools). Secure the memory (data).

In the Part 7, we’ll show the full end-to-end traffic flow diagram that makes “north–south” and “east–west” enforcement intuitive.

Recent Posts













Gartner Research Report

2025 Gartner® Magic Quadrant™ for SASE Platforms

Versa has for the third consecutive year been recognized in the Gartner Magic Quadrant for SASE Platforms and is one of 11 vendors included in this year's report.