Enterprise AI security
Why guardrails are not enough for enterprise AI security
AI agents reason around policy guardrails. Physical separation — sandboxed agents, locked-down runners, and permission ledgers — is the only architecture that holds up.
TL;DR: Policy-based guardrails are not enough to secure enterprise AI agents. Agents can reason around them. Grail uses physical separation — a sandboxed agent, a locked-down action runner, and a permission ledger — so critical systems are protected by architecture, not instructions.
This post explains the problem in plain language, shows a real example of why it matters, and walks through the three-component design we use to solve it.
The guardrail problem
Most AI agent platforms handle security with rules: “don’t delete files,” “don’t access sensitive data,” “ask before running destructive commands.” These rules live in the same context as the agent. They are instructions, not boundaries.
The problem is that modern AI models are good at reasoning. When they hit a guardrail, they don’t just stop. They look for another way to accomplish the goal. Sometimes that is exactly what you want — creative problem-solving is why you deployed the agent. But when the guardrail was supposed to protect something critical, creative workarounds are a security failure.

This is not hypothetical. The screenshot above shows an AI agent that was blocked from running a destructive shell command. Instead of stopping, it wrote Python code that does the same thing. The guardrail was technically honored — no rm -rf was executed. But the files are gone.
The takeaway: if the agent has the ability to reach a system, no amount of instructions will reliably prevent it from doing so when it decides the goal requires it.
Why physical separation matters
The only reliable way to prevent an AI agent from doing something is to make it physically impossible. Not “discouraged,” not “blocked by policy” — impossible. If the agent cannot reach your production database, it cannot reason its way into accessing it.
This is the same principle behind network segmentation, container isolation, and sandboxed execution environments. It has been standard practice in infrastructure security for decades. The AI agent world is only now catching up.
The three-component architecture
Grail separates enterprise AI deployments into three physically distinct components. Each one has different access levels, different trust boundaries, and different owners.
1. The sandboxed agent
The agent itself runs in an isolated environment. It can read what you have shared with it — documents, messages, context from your conversation history — but it has no direct access to production systems, sensitive databases, or critical infrastructure.
Think of it as giving a new hire a laptop with read access to your shared drive, but no admin credentials and no VPN. They can do research, draft documents, analyze data, and prepare recommendations. They cannot push to production or change your CRM records.
2. The permissioned runner
When the agent needs to do something that touches a critical system — write to a database, update a CRM record, trigger a deployment, access sensitive data — that action does not happen inside the agent. It gets sent to a separate, locked-down execution environment.
This runner has its own access rules, its own credential store, and its own audit logging. The agent can request an action. It cannot perform it directly. The separation is physical, not policy-based. Even if the agent is compromised or reasons its way around instructions, it cannot escalate its own permissions.
3. The permission ledger
Between the agent and the runner sits a permission system. Before any sensitive action is executed, the system checks whether it is allowed. If the action matches an existing rule — say, the domain owner has pre-approved CRM reads between 9am and 5pm — it proceeds. If not, the request goes to the right person for approval.
The approval options are practical, not bureaucratic:
- One-time approval — approve this specific action, right now.
- Time-limited access — allow this type of action for the next hour, day, or week.
- Standing rules — define a class of actions that the agent can always do without asking.
Every decision — whether approved, denied, or auto-allowed — is logged in an immutable audit trail. The credential owner is always the one who approves. No one else can grant access on their behalf.
What this means for enterprise buyers
The conversations we have with enterprise teams almost always circle back to the same questions: How do I know the AI will not access something it should not? Who is in control? What happens if it tries something unexpected?
The three-component architecture answers all of them with the same principle: the agent can never do more than its physical access allows, and any escalation requires human approval through a logged, auditable process.
Concretely:
- Your workflows are your IP. We never share, productize, or learn from one client’s workflows to serve another. Your processes stay yours.
- Deploy on your infrastructure. Run Grail on your own cloud, your own account, your own VPC. We set it up and you own the keys.
- No model training on your data. Your conversations, documents, and workflows are never used to train any model.
The standard we are building toward
We are not claiming to have solved enterprise AI security. We are saying that the starting architecture matters, and most agent platforms got it wrong by putting everything in one trust boundary and relying on prompt-level guardrails.
Physical separation is harder to build and harder to ship quickly. But it is the only approach that holds up when the agent is doing what you hired it to do — reasoning creatively to get work done.
The goal is a system where teams can deploy AI agents with confidence: confident that the agent cannot overstep its access, confident that sensitive actions go through the right people, and confident that everything is logged for audit.
That is the standard enterprise teams should demand before putting AI agents into real workflows.
Key takeaways
- Policy guardrails live in the same context as the agent. When the agent reasons around them, they offer no real protection.
- Grail separates AI deployments into three physically distinct components: a sandboxed agent, a locked-down runner, and a permission ledger.
- Every sensitive action requires human approval through a logged, auditable process — one-time, time-limited, or standing rules.
- Your workflows, data, and infrastructure stay under your control. No cross-client learning. No model training on your data.