
Building what NIST is asking for: an AI agent security proxy
NIST just closed a public RFI on AI agent security. The question they were asking, in five different ways: how do you constrain what an AI agent can do, and how do you prove it was constrained? We built something that answers both. Not because we read the RFI — we built it because we ran into the problem first. Reading the RFI afterward was like seeing someone formally describe a thing you've been fixing with duct tape. The problem frameworks don't solve Most security frameworks for AI agents focus on what the agent should do: don't call dangerous APIs, don't exfiltrate data, follow least-privilege principles. Good policies. But policies are enforced at configuration time, and AI agents operate at runtime. The gap between "the policy says X" and "the agent did X" is where incidents happen. The deeper issue: in heterogeneous pipelines, each provider certifies only their own model. AWS certifies Bedrock. OpenAI certifies GPT-4o. Your self-hosted Mistral is self-attested at best. The hand
Continue reading on Dev.to
Opens in a new tab

