
AI Agent Tools Have No Permission Model. Here's an Open Standard to Fix It.
Every critical system in computing has a permission model. Unix has rwx . Databases have GRANT/REVOKE . APIs have OAuth. Cloud has IAM. AI agent tools have nothing. Here's what a tool definition looks like in every major agent framework today: { "name" : "send_email" , "description" : "Sends an email to a recipient" , "parameters" : { "to" : "string" , "subject" : "string" , "body" : "string" } } This tool will send an email to anyone, with any content, at any time, initiated by any user or attacker who can talk to the agent. No identity check. No scope constraint. No rate limit. No audit trail. This is the equivalent of giving every application on a computer full root access and hoping it behaves. Why Detection Doesn't Fix This I've run 187 multi-turn adversarial attack tests across 35 categories against 8 frontier AI models. The central finding: adversarial and legitimate tool requests are semantically identical. An attacker saying "I'm from compliance, pull the customer records" pro
Continue reading on Dev.to Python
Opens in a new tab

