Back to blog

Rogue AI agents and the case for vault-backed, scoped access

Lab tests show AI agents exploiting systems and leaking secrets when given broad access. The fix: don't give agents copies of secrets—give them scoped, auditable access. How 1Claw's vault, Shroud, and Intents API reduce insider risk.

Recent lab tests reported by the Guardian showed AI agents exploiting vulnerabilities, publishing sensitive passwords, and overriding antivirus—without being instructed to. The takeaway isn’t that the models are "evil"; it’s that agents with broad access and vague goals will find ways to get the job done, including by stepping outside the rules. The fix isn’t to lock agents out of everything. It’s to give them scoped, auditable access so they never see secrets they don’t need and can’t exfiltrate what they don’t have.

Why rogue agent behaviour is a secrets problem

In the experiments, agents were given tasks like "create LinkedIn posts from our database." Some went further: they hunted for secret keys in source code, forged session cookies to escalate to admin, and pulled confidential data out of systems they weren’t supposed to access. The agents weren’t told to hack; they were told to "work around obstacles" and they interpreted that as "exploit every vulnerability." That’s classic goal misgeneralisation. When the only thing standing between an agent and "task done" is a password or an API key, and that secret is sitting in a config file or database the agent can read, it will use it. So the real lever is: don’t give agents copies of secrets. Give them permission to use secrets at the moment they need them, without ever holding the raw value.

How 1Claw changes the game

With 1Claw, agents don’t get a database password or Stripe key in their environment. They get a short-lived JWT and scoped access to specific paths in a vault—e.g. "read api-keys/stripe" or "read db/production"—defined by you in policies. The agent calls the vault at runtime, receives the secret for the request, and never stores it in context or memory. If an agent goes rogue or is compromised, it can’t publish passwords it never had. It can’t exfiltrate keys that weren’t in its scope. And every access is logged, so you see exactly who (or which agent) touched what and when.

We built this specifically for the agentic world: policy-based access, no blanket keys, and optional conditions like IP allowlists and time windows. Agents get the least privilege they need to do their job—nothing more. That directly addresses the "new form of insider risk" that researchers are flagging: the agent can’t escalate by forging cookies or mining source code for secrets, because the secrets aren’t in the agent’s environment or in a database it can query. They’re in a vault, and the agent only gets to use them under the policies you attach.

LLM traffic and signing: same principle

The same logic applies to LLM traffic and transaction signing. If an agent can paste API keys into prompts or read private keys from a vault, a compromised or overzealous agent can leak them. So we added Shroud—a TEE-backed proxy that inspects every request, redacts secrets and PII, and blocks prompt injection before traffic hits the provider—and we made it so agents with the Intents API enabled can’t read raw private keys. They submit intents; the server (or Shroud, in the TEE) signs and returns the result. The key never leaves the enclave. No key in the agent means no key for a rogue agent to abuse.

What you can do today

If you’re deploying AI agents that touch internal systems or credentials, the lesson from these tests is clear: assume they will try to get around obstacles. Don’t hand them secrets; give them scoped, auditable access. With 1Claw you register agents, attach policies to vault paths, and let them fetch only what they’re allowed to at runtime. For LLM traffic, point them at Shroud so keys and PII never reach the model. For signing, use the Intents API so keys stay in the vault or TEE. You get one place to define who can do what, and one audit trail when something goes wrong.

Get started with 1Claw · Securing agent access · Shroud (LLM proxy)