Skip to content

Codex CLI Overview (Beta)

Codex CLI daily report

⚠️ The Codex companion CLI is experimental. Expect breaking changes while both ccusage and OpenAI's Codex CLI continue to evolve.

The @ccusage/codex package reuses ccusage's responsive tables, pricing cache, and token accounting to analyze OpenAI Codex CLI session logs.

Installation & Launch

bash
# Recommended - always include @latest
npx @ccusage/codex@latest --help
bunx @ccusage/codex@latest --help  # ⚠️ MUST include @latest with bunx

# Alternative package runners
pnpm dlx @ccusage/codex --help
pnpx @ccusage/codex --help

# Using deno (with security flags)
deno run -E -R=$HOME/.codex/ -S=homedir -N='raw.githubusercontent.com:443' npm:@ccusage/codex@latest --help

⚠️ Critical for bunx users

Bun 1.2.x's bunx prioritizes binaries matching the package name suffix when given a scoped package. For @ccusage/codex, it looks for a codex binary in PATH first. If you have an existing codex command installed (e.g., GitHub Copilot's codex), that will be executed instead. Always use bunx @ccusage/codex@latest with the version tag to force bunx to fetch and run the correct package.

Since npx @ccusage/codex@latest is quite long to type repeatedly, we strongly recommend setting up a shell alias for convenience:

bash
# bash/zsh: alias ccusage-codex='bunx @ccusage/codex@latest'
# fish:     alias ccusage-codex 'bunx @ccusage/codex@latest'

# Then simply run:
ccusage-codex daily
ccusage-codex monthly --json

TIP

After adding the alias to your shell config file (.bashrc, .zshrc, or config.fish), restart your shell or run source on the config file to apply the changes.

Data Source

The CLI reads Codex session JSONL files located under CODEX_HOME (defaults to ~/.codex). Each file represents a single Codex CLI session and contains running token totals that the tool converts into per-day or per-month deltas.

What Gets Calculated

  • Token deltas – Each event_msg with payload.type === "token_count" reports cumulative totals. The CLI subtracts the previous totals to recover per-turn token usage (input, cached input, output, reasoning, total).
  • Per-model grouping – The turn_context metadata specifies the active model. We aggregate tokens per day/month and per model. Sessions lacking model metadata (seen in early September 2025 builds) are skipped.
  • Pricing – Rates come from LiteLLM's pricing dataset via the shared LiteLLMPricingFetcher. Aliases such as gpt-5-codex map to canonical entries (gpt-5) so cost calculations remain accurate.
  • Legacy fallback – Early September 2025 logs that never recorded turn_context metadata are still included; the CLI assumes gpt-5 for pricing so you can review the tokens even though the model tag is missing (the JSON output also marks these rows with "isFallback": true).
  • Cost formula – Non-cached input uses the standard input price; cached input uses the cache-read price (falling back to the input price when missing); and output tokens are billed at the output price. All prices are per million tokens. Reasoning tokens may be shown for reference, but they are part of the output charge and are not billed separately.
  • Totals and reports – Daily, monthly, and session commands display per-model breakdowns, overall totals, and optional JSON for automation.

Environment Variables

VariableDescription
CODEX_HOMEOverride the root directory containing Codex session folders
LOG_LEVELAdjust consola verbosity (0 silent … 5 trace)

When Codex emits a model alias (for example gpt-5-codex), the CLI automatically resolves it to the canonical LiteLLM pricing entry. No manual override is needed.

Next Steps

Have feedback or ideas? Open an issue so we can improve the beta.

Troubleshooting

Why are there no entries before September 2025?

OpenAI's Codex CLI started emitting token_count events in commit 0269096 (2025-09-06). Earlier session logs simply don't contain token usage metrics, so @ccusage/codex has nothing to aggregate. If you need historic data, rerun those sessions after that Codex update.

What if some September 2025 sessions still get skipped?

During the 2025-09 rollouts a few Codex builds emitted token_count events without the matching turn_context metadata, so the CLI could not determine which model generated the tokens. Those entries are ignored to avoid mispriced reports. If you encounter this, relaunch the Codex CLI to generate fresh logs—the current builds restore the missing metadata.

Released under the MIT License.