Skip to content

Logged Usage

Logged Usage is for learning how to run sessions more efficiently — not accounting. Use it to understand which tasks, prompts, and workflows cost the most so you can improve over time.

The figures shown are estimates derived from the JSONL session files Claude Code writes to disk. Some API costs are not written to those files. For accurate billing, always refer to your Anthropic Console or subscription dashboard. Anthropic is the sole source of truth for your actual spend.

Cate reads Claude Code’s local JSONL session files and aggregates token counts into cost estimates using Anthropic’s published retail API pricing. It applies these estimates at the issue level, so you can see what a specific ticket or feature cost — not just a single chat.

What makes this different from other Claude Code cost trackers:

  • Issue-level aggregation. Costs roll up across all sessions tied to a Jira, Linear, or GitHub issue — planning, coding, review, and pairing sessions combined. This is the first tool to track spend at the ticket level.
  • No double-counting of batched messages. A common error in other trackers; Cate corrects for it.

Estimates use retail API key pricing by default. If you’re on Claude Pro, Max, or an enterprise agreement, your actual cost may be zero or significantly different. See Settings → General to estimate your discount.

Click Usage in the left sidebar to open the project-level usage view.

Four cards at the top give a quick read on the selected time window (1d / 7d / 30d / Custom):

CardWhat it shows
Logged UsageTotal estimated cost across all sessions in the window
Logged Usage / IssueAverage cost per issue — useful for spotting outliers
IssuesNumber of issues with at least one logged session
PRsNumber of pull requests opened

A stacked bar chart breaking down spend by day and by model. Each bar segment corresponds to a model (e.g. haiku-4, opus-4, sonnet-4). Use this to spot expensive days and correlate them with the work that happened.

A ranked list of the most expensive issues in the time window. Click any row to expand it and see the individual sessions that contributed to that cost.

A stacked area chart showing how time was spent across workflow phases: Planning, AI Coding, AI Review, Reviewing, and Pairing. This is useful for understanding where sessions are running long relative to the value delivered.

The table below the charts lists every issue with logged usage. Each row shows:

  • PR — whether Cate opened a PR for this issue
  • Issue — the issue ID and title
  • Est. Cost — total estimated cost for that issue
  • In / Out — input and output token counts
  • CW / CR — cache write and cache read tokens
  • Duration — total wall-clock time across all sessions
  • Last Activity — when the most recent session ran

Expand any issue row to see the individual sessions that made it up, broken down by workflow phase (Planning, AI Coding, AI Review, Reviewing).


While inside any agent session, click Usage in the top toolbar to open the session-level usage panel.

This view refreshes every 30 seconds. Enable Tail (top right of the panel) to refresh every 5 seconds.

CardWhat it shows
Total TokensCombined input + output + cache tokens for this session
CacheCache write and cache read token counts
Hit RatePercentage of input tokens served from cache
Est. CostEstimated cost, adjusted by your discount factor from Settings → General

Cache hit rate — the most important metric

Section titled “Cache hit rate — the most important metric”

A high cache hit rate (80%+) means most of your input tokens are being served from Anthropic’s prompt cache rather than billed at full input rates. Cache reads cost roughly 10× less than standard input tokens.

If your hit rate is low, common causes:

  • The context window is being cleared between requests (compaction triggered too aggressively)
  • The system prompt or large context blocks are changing between turns, breaking cache keys
  • Sessions are very short — the cache warms up over multiple turns

A chart showing token counts per request, broken down by Cache Read, Cache Write, Input, and Output. The red line overlays the running cache hit rate. Use this to watch the cache warm up over the course of a session and to spot requests that are unusually expensive.

A per-request breakdown of every API call in the session. Columns:

ColumnDescription
#Request sequence number
In / OutInput and output tokens for this request
CW / CRCache write and cache read tokens
CostEstimated cost for this request
Hit%Cache hit rate for this request
Total CostRunning cumulative cost

The goal of Logged Usage is to help you develop better practices around how agent sessions consume tokens — so you can write tighter tasks, structure context better, and avoid waste.

A few patterns worth watching for:

  • High cost per issue with low PR count — tasks may be too large or too vague. Smaller, well-scoped issues tend to be more token-efficient.
  • Low cache hit rate in coding sessions — consider whether your CLAUDE.md or system prompt is stable across turns. Frequent changes break cache keys.
  • Planning sessions that cost as much as coding sessions — a spec that requires many revision cycles may need a clearer initial prompt.
  • Expensive AI Review sessions — review agents re-read the full spec and diff. Very large PRs drive up review cost; stacked PRs (via Epics) help keep individual PR size manageable.