Headroom

Claude Code usage: what counts and how to get more from your plan

If your problem is plan limits rather than billing, this is the page to read. Claude Code usage is driven by how much context the model has to read and write across a session.

What counts toward Claude Code usage?

Usage is driven by the total tokens Claude Code reads and writes during a session. That includes your prompt, files it opens, terminal output, logs, search results, JSON payloads, tool responses, and the model output it sends back.

Which workflows burn usage fastest?

Debugging from logs, codebase exploration, large JSON payloads, shell-heavy tasks, and repeated file reads are usually the worst offenders. These workflows pile on machine-generated context quickly, so a few long sessions can chew through a plan much faster than expected.

How do you get more from your plan?

The best approach is to reduce noise before it reaches the model. Keep the useful context, strip the repetitive stuff, and make verbose inputs smaller before Claude Code reads them. That is where tools like RTK, Distill, MemStack, and Headroom help most.

If you want the workflow details, read the Claude Code cost guide and compare the benchmark section against the kinds of sessions your team runs most.

How is Claude Code usage different from Claude Code cost?

Usage is about how much of your plan you consume. Cost is about the token spend behind that usage. The two are related, but not identical: the same noisy session that burns through your plan faster also tends to make Claude Code feel more expensive.

Make your Claude plan last longer

Download the app, run one of your normal Claude Code sessions, and compare the tokens before and after optimization.

Download for free

macOS only ยท Linux coming soon