Your Claude Code fleet retried the same six errors
1,200 times last week.
Fukura shows you which fixes actually worked, ranked by how much engineer time they saved. Every failed tool call from Claude Code, Cursor, or your own scripts is captured locally, redacted, and measured against its outcome — so the next agent doesn’t pay the same cost twice.
curl -sSL https://fukura.dev/install.sh | bashAlso available: cargo install fukura, or read the script first at /install.sh.
Agents bang on tools. Nothing remembers.
A coding agent runs a hundred shell commands to land one pull request. Ninety of them fail. The ten fixes that worked are thrown away the moment the session ends — next week, a different agent on a different branch retries the same wrong command, hits the same Terraform error, and burns the same forty-five minutes.
Sentry sees production crashes. Datadog sees metrics. Your editor forgets everything. Nobody is recording the intermediate failure-and-fix loop that now dominates how software gets written. That loop is where the cost lives, and it’s invisible.
How it works
Fukura is a local-first CLI and an open spec (EKP). The CLI watches your shells and agent sessions, captures error/fix pairs, redacts secrets against deterministic rules, and stores them in a content-addressable repo you own. A hub — optional, self-host or managed — aggregates across machines so your whole team queries the same memory.
Shell hooks (zsh / bash / fish / powershell) + MCP tools catch every failing invocation. Redaction runs before anything leaves your machine.
EKP adapters turn raw stderr into a stable fingerprint so the same error from any tool, language, or agent collapses to the same key.
Every attempt at a fix is recorded with its outcome. The hub surfaces success rates per fingerprint, human vs. agent splits, and unresolved patterns in real time.
You can’t improve a loop you can’t measure.
Fukura’s effectiveness score — (successful next command ÷ attempts) weighted by time-to-fix — is the first number that tells you whether your agent tooling is getting better or just getting louder. Ship it to a dashboard. Put it on a review. Sort your docs by it. The hub does the arithmetic; you decide what to do with the answer.
Ranked by attempts × (1 − success rate). The concrete list every DevEx standup should end with.
Which fingerprints do agents disproportionately fail on? That’s where a one-paragraph note pays back the most.
Percent of active fingerprints that have at least one team note. The single number that says “how much of our firefighting is actually captured”.
Install in one line. Self-host the hub when you’re ready.
The CLI is Apache-2.0 and runs offline. The hub is source-available commercial — free for individuals and small teams, paid past that. You can turn it on whenever the effectiveness data starts to matter.