Context optimization for Claude Code

Your agent is wasting context.

Optiforce strips the fluff before it reaches Claude — old tool outputs, stale history, and bloated thinking blocks that pile up with every agent step. Same results. 40% fewer tokens.

$npx optiforce setupclick to copy
Context per agent turn — 103 baseline vs 65 gateway SWE-bench trialsBenchmark data
Without Optiforce
With Optiforce
The problem

Context grows.
Relevance doesn't.

Claude Code appends every tool output and result to the context window with no pruning. By step 10, the model is reading your step 1 bash output. That's your token budget.

01

Every step adds more

Each agent iteration appends tool outputs, results, and reasoning — without any pruning or relevance check.

02

Old context stays forever

Early steps that no longer matter remain in the prompt, consuming tokens that could go toward what's relevant now.

03

Signal gets buried

As context accumulates, the model's attention spreads thin over outdated content instead of focusing on the current task.

04

You pay for the noise

Every redundant token costs money and burns rate-limit capacity — regardless of whether it contributes to the output.

How it works

Three steps. Zero config.

01

Intercept

Optiforce runs as a lightweight HTTPS proxy alongside Claude Code. No changes to your code or workflows.

02

Rank & Filter

Each request is analyzed in real-time. Prior context is ranked by relevance to the current step. Low-signal history is compressed out.

03

Forward

A focused, high-signal prompt reaches Anthropic's API. Context stays bounded across the entire agent run — automatically.

Real Data

Actual savings
by command type.

Per-command savings measured across 13,000+ real compactions. Every bar is a real tool call — left is what Claude Code sent, right is what Optiforce forwarded.

Without Optiforce — full output sent
  • Context grows with every step
  • No prioritization of past information
  • Increasing dilution over long runs
  • Every turn makes the model read more
With Optiforce — signal only
  • Context stays bounded and relevant
  • Only high-signal steps are included
  • Stable efficiency across long runs
  • Predictable, controlled token usage
Benchmark

Fewer tokens. Same quality.

Head-to-head on SWE-bench Verified against RTK v0.35.0 — the leading open-source context compressor. Optiforce cuts more tokens while matching or exceeding task quality.

TaskRTK v0.35.0Baseline (no compression)Optiforce
astropy1.98M tokens· reward 0.601.84M tokens· reward 0.851.06M tokens· reward 0.85-47% vs RTK
django2.71M tokens· reward 0.952.21M tokens· reward 0.911.38M tokens· reward 0.91-49% vs RTK
matplotlib1.68M tokens· reward 0.951.65M tokens· reward 1.000.93M tokens· reward 1.00-45% vs RTK
384M
Characters stripped from
context since April 2026
60%
Average savings per
compacted tool call
13K
Compactions processed
across 8,860 agent sessions
Who it's for

Built for every
Claude Code workflow.

Individual

Developer

Stay focused during long Claude Code runs. Plug in once and immediately get more from every session — no configuration needed.

  • Plug-and-play, zero config
  • Better efficiency in everyday workflows
  • Free tier available
Teams

Engineering Teams

Predictable token usage, consistent agent behavior, and full observability into context selection — across your entire organization.

  • Predictable token usage at scale
  • Observability into context selection
  • Designed for production workflows
  • Consistent behavior across team runs
Trust & Privacy

Your code stays yours.
We just count the signal.

No code stored

Your source code and context content are never stored. Only anonymized per-command-type statistics are retained temporarily for your dashboard.

Stats with a short shelf life

Aggregated usage statistics — which command types waste context, how much was saved — are kept for a limited time to power your dashboard, then discarded.

Zero code changes

Works as a transparent proxy alongside Claude Code. No modifications to your codebase, workflows, or tooling.

Lightweight layer

An external optimization layer that adds minimal latency and delivers maximum signal improvement at every step.

Send less.
Get more.

Start optimizing your Claude Code context today. Free for individuals.