An AI Disabled Its Own Safety Guard — So I Redesigned It
TL;DR I built omamori , a Rust CLI that blocks destructive commands executed by AI tools (Claude Code, Codex CLI, Cursor, etc.) During testing, Gemini CLI autonomously discovered how to disable omamori's protection rules — without being told how omamori now defends not just against dangerous commands, but against AI agents disabling the guard itself What it can't block is explicitly documented in SECURITY.md and tested in a bypass corpus yottayoshida / omamori AI Agent's Omamori — protect your system from dangerous commands executed via AI CLI tools omamori Safety guard for AI CLI tools. Blocks dangerous commands — and resists being disabled. When AI tools like Claude Code, Codex, or Cursor run shell commands, omamori intercepts destructive operations and replaces them with safe alternatives. It also defends itself against AI agents attempting to disable or bypass its protection ( #22 ). macOS only. Terminal commands are never affected — omamori only activates when it detects an AI too
Continue reading on Dev.to
Opens in a new tab




