
Prompt Versioning in Production: What We Learned Running LLM Agents for 3 Months
Our SDR agent's system prompt went through seven iterations before it stopped guessing email addresses. Here is what that process taught us about treating prompts as production code. We run six AI agents in production, daily, on an automated schedule. Each agent has a system prompt stored as a markdown file in a git repository. Over three months, those prompts have accumulated more commits than most of our Python scripts. The prompts are the most frequently edited files in the codebase. This was not what we expected. We expected to write a prompt, tune it for a week, and leave it alone. What actually happened is that prompts behave like code: they have bugs, they need tests, they regress when you change them, and they require review before deploying to production. The tooling and practices around software engineering apply directly. Here is what we learned. Prompts Are Markdown Files in Git Each agent's system prompt lives in .claude/agents/{agent-name}.md . The CMO agent has cmo.md .
Continue reading on Dev.to Python
Opens in a new tab

