Token burn
An informal but increasingly load-bearing piece of vocabulary among practitioners running agentic AI in production.
In one sentence
Token burn is the colloquial term for the rate at which an agent is consuming model tokens (and therefore money) — usually invoked at the precise moment the practitioner realizes the rate is much higher than expected.
In one slightly longer sentence
Token burn is the rate at which a configured-with-good-intentions agentic system silently transmutes electricity, attention, and credit-card balance into JSON, while the operator sleeps soundly believing the heartbeat-every-30-minutes was a thrifty design choice.
Why this term exists
Cloud-hosted frontier models charge per token of input and output. A casual chat costs cents. A heavily-instrumented agent — heartbeats every 30 minutes, sub-agents firing in parallel, large context windows reloaded every turn, multiple tools each producing verbose output — does not cost cents.
The arithmetic is unforgiving. Forty-eight heartbeats per day, each loading 100,000 tokens of context, against a frontier model at $15-30 per million tokens, runs to real money in a hurry. Practitioners discovered this collectively in 2025-2026 when the first generation of always-on agentic systems shipped and the first round of monthly bills landed.
The phrase token burn was the natural label. It rhymes spiritually with cash burn from the startup vocabulary, with the same connotation: a continuous expenditure that goes on whether you are watching it or not, and whose true rate you discover only by inspecting the receipts.
A taxonomy of token-burn modes (informal)
Practitioner experience has produced an informal classification:
- Slow burn. The default state of a configured-but-undermonitored agent. Steady, low-grade consumption. Easy to live with for weeks before anyone notices.
- Heartbeat burn. Dominant after the first few weeks. Periodic checks compound into the largest single line item on the bill. Cured by tiering the heartbeat model to something cheaper or local.
- Sub-agent burn. Less visible than heartbeat burn because sub-agents fire only when something happens — but each spawn loads the parent context plus its own working memory. A misconfigured sub-agent loop has been known to consume a thousand-dollar credit balance in an afternoon.
- Compaction burn. When an agent’s context fills up and the runtime calls a compaction model to summarize and shrink it, the compaction itself is a model call. On a chatty agent this is recurring overhead the operator usually forgets exists.
- Loop burn. The catastrophic mode. An agent enters a self-stimulating loop — calling a tool, reacting to its output, calling the tool again, reacting again — and burns its way through the operator’s monthly budget before anyone wakes up. The phrase “I went to bed with $300 of credit and woke up to a $2 budget alert” belongs to this category.
Working example from this machine
The OpenClaw Control dashboard on this MacBook reports cumulative cost in plain numbers. As of the morning of May 2, 2026, 30 million tokens have flowed through the gateway across 351 messages — a number perfectly explicable by the workload of a real lecturer running real classes through a real agent for real weeks, but a number whose existence is itself a small education in what production agentic AI actually costs.
The model-tiering plan filed the same morning is, structurally, a token-burn-management plan. The vocabulary moved from finance (“how much are we spending?”) to ecology (“which fires are we letting burn, which are we putting out, and which are we converting to controlled cooking flames on the local hardware?”). The shift in metaphor is, itself, a small zhengming — see the naming entry.
Why this matters in a teaching context
For a BBA or MBA classroom, token burn is a useful entry point into a much older management concept dressed in fresh terminology: the difference between the cost you intend to incur and the cost you are actually incurring.
Every accounting student learns about variance analysis. Every operations course covers process drift. Every strategy course mentions the McKinsey adage about budgets being descriptions of past intentions rather than future behaviour. Token burn is all of these rolled together in a domain where the gap between intention and behaviour is unusually fast — billable seconds rather than billable months.
A useful classroom exercise: take a published agentic-system architecture diagram (most vendor decks have one) and ask students to estimate where the token burn lives. The answers reveal who has read about the technology and who has actually run it.
How practitioners diagnose it
The standard moves, in order of how angry the practitioner has become:
- Look at the dashboard. Most agentic platforms surface monthly cost. Many practitioners discover the existence of this dashboard only after their first surprise bill.
- Review heartbeat frequency. Halve it. Watch the cost line. Halve it again if needed.
- Tier the models. Frontier model for the work that needs it; smaller / cheaper / local model for everything else. (See Ollama, the model-tiering pattern, and the practical phased-rollout approach.)
- Check for loops. Look for any agent that calls a tool, reads the tool’s output back into itself, and decides — based on that output — to call another tool. If the loop has no termination condition, it is a token-burn time bomb.
- Trim the context. The same heartbeat that re-injects 100,000 tokens of workspace bootstrap every fire can be configured to inject only what it needs. The operational difference is dramatic.
- Set alerts. Most providers offer budget alerts. Configure them. Configure two. The first warns; the second cuts off.
The seasoned practitioner’s stages of grief
In rough order:
- Denial. “Surely this is just the trial period.”
- Anger. “Why is the heartbeat using Opus?”
- Bargaining. “What if I run heartbeats only during business hours?”
- Depression. “I built a system that pays a frontier model to ask itself if anything has happened in the last 30 minutes.”
- Acceptance. “Time to tier.”
Most practitioners reach acceptance within one billing cycle. The truly fortunate reach it before their first billing cycle, by reading articles like this one.
Trade-offs
- Aggressive cost cutting can break behaviour. A heartbeat downgraded too far becomes useless. A sub-agent on too small a model misses the nuance the user needed. Token burn management is genuinely a trade-off curve, not a free optimization.
- Local models change the math but do not eliminate it. Local inference costs are electricity, hardware depreciation, and your time, not zero. The bills look smaller because you absorb them differently, not because they vanished.
- Monitoring is the foundation. Without good per-job, per-tier, per-day cost telemetry, every cost optimization is a guess. The dashboard exists for this reason. Look at it more often than feels necessary.
Related and adjacent terms
- Cash burn (from startup finance) — the mother metaphor.
- Compute burn (general, predates LLMs) — the same shape applied to GPU hours, container minutes, or any metered cloud resource.
- Context burn (less common) — specifically the cost of repeatedly re-loading large context windows.
- Wallet burn (humorous, internal) — what spouses of heavy practitioners call the same phenomenon, with no further taxonomy needed.
Related entries: Heartbeat, Sub-agent, Ollama. For the architectural escape hatch, see the model-tiering plan (forthcoming as architecture/model-tiering.md).
Filed in the spirit in which most working terminology is filed: from the position of someone who learned the term the expensive way.