
Counting Bullets: Why Token Burn Is the Wrong Metric for Agent Work
Meta and OpenAI are running internal leaderboards. Not for commits shipped, or bugs fixed, or products launched. For tokens consumed . One OpenAI engineer reportedly burned through 210 billion tokens in a single week — the equivalent of reading 33 Wikipedias, processed and discarded. This is apparently now a performance metric worth tracking. The phenomenon has a name: tokenmaxxing . Gizmodo called it like telling soldiers to gauge their battlefield success by the number of bullets fired. They're right. But the real problem is subtler, and it cuts deeper into how we're thinking about agent productivity in 2026. The Overhead Problem Tyler Folkman ran an experiment this week. He asked a simple question — "what are the 3 largest cities in Utah?" — through a raw API call and through a modern agentic framework (LangChain's Deep Agents). Raw API: 77 tokens. Through the agent framework: 5,983 tokens. Seven LLM calls. A 78x multiplier. For a more complex task — a bug fix requiring file reads a
Continue reading on Dev.to
Opens in a new tab
