savethetokens

Reduce Claude Code token burn with proactive compacting, task-scoped sessions, checkpoint files, tiered docs, and relevance-based context pruning. Use when a user explicitly asks to reduce context usage/cost, run token benchmarks, or optimize CLAUDE.md. Do not use to expand product scope or skip correctness checks.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "savethetokens" with this command: npx skills add redclawww/savethetokens/redclawww-savethetokens-savethetokens

Context Governor

Optimize context usage with practical, high-impact workflows and scripts.

Non-Negotiable Guardrails

  1. Keep scope locked to the user request. Do not add extra features, pages, or telemetry unless asked.
  2. Treat token optimization as a constraint, not the goal. Correctness and security win over token reduction.
  3. Never claim token savings without before/after measurement on comparable tasks.
  4. If context-saving actions risk quality loss, keep the extra context and state the tradeoff.
  5. Never reduce code thoroughness to save tokens. Tests, strict config, safety checks, and error handling are non-negotiable. Save tokens from message verbosity — never from output completeness.

Operating Modes

  • Lean Mode (default): Use lightweight context hygiene only; do not create new benchmark artifacts.
  • Measurement Mode: Use launch-readiness or A/B telemetry scripts only when user asks for proof/percentages.

Claude Code Message Budget (required)

  1. Keep progress updates short and phase-based. Do not narrate every file write.
  2. Do not paste long command output unless user asks. Summarize only key signals.
  3. Do not repeat the same command without a code/input change; if retried, state the reason once.
  4. If /context shows message growth is unusually high, switch to stricter concise mode:
    • fewer updates
    • shorter summaries
    • batch related edits before reporting
  5. Prefer one concise final summary over long running commentary.
  6. For benchmark runs, enforce matched behavior on both variants:
    • same stop criteria
    • same compact policy
    • same output style (no extra giant report in one variant only)

Operating Playbook

  1. Confirm objective and lock scope in one sentence.
  2. Keep one chat session per task. Start a new session for unrelated work.
  3. Use ! <command> for direct shell commands when no reasoning is required.
  4. Run /context periodically. Compact around 50% usage instead of waiting for hard limits.
  5. Before /compact or /clear, create a checkpoint file with next steps and touched files.
  6. Keep top-level docs lean; move deep details to linked docs/*.md.
  7. Before final output on code tasks, run the quality gates in docs/QUALITY_GATES.md.
  8. For token-savings claims, run matched A/B using docs/BENCHMARK_PROTOCOL.md.
  9. For Claude benchmark runs, use docs/STRICT_BENCHMARK_PROMPT.md as the session starter.

Quick Commands

# Generate execution plan
python ~/.claude/skills/savethetokens/scripts/govern.py --budget 8000

# Generate checkpoint before compact/clear
python ~/.claude/skills/savethetokens/scripts/session_checkpoint.py \
  --task "..." \
  --done "..." \
  --next "..." \
  --context-percent 52 \
  --message-count 36

# Create session hook (Claude Code)
python ~/.claude/skills/savethetokens/scripts/session_hook_generator.py --project .

# Optimize CLAUDE.md
python ~/.claude/skills/savethetokens/scripts/claude_md_optimizer.py --analyze

# Calculate cost savings
python ~/.claude/skills/savethetokens/scripts/cost_calculator.py --developers 5

# Run launch-readiness benchmark with section-wise savings
python ~/.claude/skills/savethetokens/scripts/launch_readiness.py

# Run live A/B telemetry session (auto split control/optimized)
python ~/.claude/skills/savethetokens/scripts/govern.py \
  --input context.json \
  --budget 8000 \
  --experiment-id claude-launch-v1 \
  --variant auto \
  --assignment-key TICKET-123

# Generate measured A/B report from live sessions
python ~/.claude/skills/savethetokens/scripts/ab_telemetry.py \
  --experiment-id claude-launch-v1 \
  --days 14 \
  --required-intents code_generation,debugging,planning,review \
  --min-samples-per-intent 5

# Strict mode: exit 2 if claim gates fail (CI-friendly)
python ~/.claude/skills/savethetokens/scripts/ab_telemetry.py \
  --experiment-id claude-launch-v1 \
  --strict-claim-mode

# Print report JSON to stdout (pipe to jq, etc.)
python ~/.claude/skills/savethetokens/scripts/ab_telemetry.py \
  --experiment-id claude-launch-v1 \
  --json-stdout

# Code-task quality gate checklist (required before final answer)
cat ~/.claude/skills/savethetokens/docs/QUALITY_GATES.md

# Compare two /context snapshots (control vs optimized)
python ~/.claude/skills/savethetokens/scripts/context_snapshot_diff.py \
  --before-file before.txt \
  --after-file after.txt \
  --strict

# Compact watchdog (advisory, safe defaults)
python ~/.claude/skills/savethetokens/scripts/compact_watchdog.py \
  --context-file context_snapshot.txt \
  --require-checkpoint

# Dynamic tool filtering (fail-open recommended)
python ~/.claude/skills/savethetokens/scripts/tool_filter.py \
  --input tools.json \
  --query "..." \
  --fail-open

# Semantic skill selection (recommendation only)
python ~/.claude/skills/savethetokens/scripts/skill_selector.py \
  --query "..."

# External memory store (bounded retrieval)
python ~/.claude/skills/savethetokens/scripts/memory_store.py search \
  --query "..." \
  --for-prompt \
  --top-k 5 \
  --max-chars 1200

# Print lean session prompt template
cat ~/.claude/skills/savethetokens/docs/LEAN_SESSION_PROMPT.md

# Print strict benchmark harness prompt
cat ~/.claude/skills/savethetokens/docs/STRICT_BENCHMARK_PROMPT.md

Scripts

ScriptPurpose
govern.pyMain entry - execution plans
analyze.pyContext analysis
prune.pyPrune to budget (max 40%)
session_hook_generator.pySession-start hooks
session_checkpoint.pySave compact-ready session checkpoints
claude_md_optimizer.pyOptimize CLAUDE.md
quick_ref_generator.pyGenerate QUICK_REF.md
tiered_context.py3-tier context classification
relevance_scorer.pyScore context relevance
cost_calculator.pyROI tracking
launch_readiness.pyLaunch benchmark + section-wise savings report
ab_telemetry.pyLive A/B telemetry report with confidence checks
context_snapshot_diff.pyDetect token regressions from /context snapshots
compact_watchdog.pySafe advisory for /compact and /clear decisions
tool_filter.pyDynamic tool filtering with fail-open safeguards
skill_selector.pySemantic skill ranking with confidence gating
memory_store.pyExternal memory store with bounded retrieval
path_filter.pyFilter package dirs

Quality Rules

  • NEVER prune system prompts, errors, recent messages
  • Max pruning: 40% (keeps quality)
  • When uncertain → KEEP content
  • Will exceed budget rather than harm quality
  • Keep solution minimal and request-aligned; avoid speculative architecture
  • Run relevant tests/checks for touched areas, or explicitly state what could not be run

Completeness Checklist (never skip under token pressure)

Token savings come from shorter messages and smarter context — never from cutting corners on output quality. Before finalizing any code task, verify:

  1. Strict config — Enable strictest compiler/linter settings available (e.g. "strict": true in tsconfig). Zero any types, zero @ts-ignore, zero @ts-nocheck.
  2. Tests for touched code — Every changed function/module has corresponding tests. Minimum: one happy path, one error path per public function.
  3. Safety limits — Runtime code that loops, recurses, or processes unbounded input must have explicit guards (max iterations, call depth, step limits, timeouts).
  4. Error handling with context — Errors include location info (file, line, span) and actionable messages. No bare catch(e) {} or except Exception: pass.
  5. Input validation at boundaries — Validate user input, API responses, and file I/O. Internal code can trust internal types.
  6. Security basics — No command injection, no unsanitized template interpolation, no hardcoded secrets. Parameterize queries.
  7. Build passes — Run type-check/compile/build before declaring done. If it can't be run, state why.
  8. State what was not verified — If any check above could not be performed (no test runner, no build script), explicitly list it in the final summary.

Where to save tokens instead: shorter progress updates, batch related edits, omit command output unless asked, compact at 50% context.

Detailed Docs (read on-demand)

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

vercel-react-best-practices

React and Next.js performance optimization guidelines from Vercel Engineering. This skill should be used when writing, reviewing, or refactoring React/Next.js code to ensure optimal performance patterns. Triggers on tasks involving React components, Next.js pages, data fetching, bundle optimization, or performance improvements.

Repository Source
216.3K23.1Kvercel
Coding

frontend-design

Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, artifacts, posters, or applications (examples include websites, landing pages, dashboards, React components, HTML/CSS layouts, or when styling/beautifying any web UI). Generates creative, polished code and UI design that avoids generic AI aesthetics.

Repository SourceNeeds Review
163.8K95.1Kanthropics
Coding

remotion-best-practices

Use this skills whenever you are dealing with Remotion code to obtain the domain-specific knowledge.

Repository SourceNeeds Review
149.9K2.2Kremotion-dev
Coding

azure-ai

Service Use When MCP Tools CLI

Repository SourceNeeds Review
137.5K156microsoft