leverage-point-audit

Audit a codebase against the 12 leverage points framework to identify gaps and improve agentic coding success.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "leverage-point-audit" with this command: npx skills add melodic-software/claude-code-plugins/melodic-software-claude-code-plugins-leverage-point-audit

Leverage Point Audit

Audit a codebase against the 12 leverage points framework to identify gaps and improve agentic coding success.

When to Use

  • Before starting a new agentic coding project

  • When agents are failing or requiring many attempts

  • When KPIs (Size, Attempts, Streak, Presence) are not improving

  • For periodic health checks of agentic capability

The 12 Leverage Points

In-Agent (Core Four)

  • Context - CLAUDE.md, README, project docs

  • Model - Appropriate model selection

  • Prompt - Clear instructions and templates

  • Tools - Required capabilities available

Through-Agent (External)

  • Standard Out - Logging for visibility

  • Types - Information Dense Keywords (IDKs)

  • Documentation - Agent-specific context

  • Tests - Self-correction capability (HIGHEST LEVERAGE)

  • Architecture - Navigable codebase structure

  • Plans - Meta-work communication

  • Templates - Reusable prompts (slash commands)

  • ADWs - Autonomous workflows

Audit Workflow

Step 1: Check Context (Leverage Points 1-4)

CLAUDE.md presence:

Search for: CLAUDE.md, .claude/CLAUDE.md Check: Does it explain the project? Conventions? Common commands?

README.md quality:

Search for: README.md Check: Does it explain structure? How to run? How to test?

Permissions configuration:

Search for: .claude/settings.json Check: Are required tools allowed?

Step 2: Check Visibility (Leverage Point 5)

Standard out patterns:

Search for: print(, console.log(, logger., logging. Check: Are success AND error cases logged? Check: Can agent see what's happening?

Anti-pattern detection:

Look for: Silent returns, bare except blocks, empty catch blocks These prevent agent visibility.

Step 3: Check Searchability (Leverage Point 6)

Type definitions:

Search for: interface, type, class, BaseModel, dataclass Check: Are names information-dense? (Good: UserAuthToken, Bad: Data)

Step 4: Check Documentation (Leverage Point 7)

Internal docs:

Search for: *.md files, docstrings, comments Check: Do they explain WHY, not just WHAT?

Step 5: Check Validation (Leverage Point 8) - HIGHEST PRIORITY

Test presence:

Search for: test_*.py, *.test.ts, *.spec.ts, *_test.go Check: Do tests exist? Are they comprehensive?

Test commands:

Check: Is there a simple test command? (npm test, pytest, etc.) Check: Do tests run quickly?

Step 6: Check Architecture (Leverage Point 9)

Entry points:

Check: Are entry points obvious? (main.py, index.ts, server.py)

File organization:

Check: Consistent structure? Related files grouped? Check: File sizes reasonable? (< 1000 lines)

Step 7: Check Templates (Leverage Point 11)

Slash commands:

Search for: .claude/commands/ Check: Are common workflows automated?

Step 8: Check ADWs (Leverage Point 12)

Automation:

Search for: GitHub Actions, hooks, triggers Check: Are workflows automated?

Output Format

After audit, provide:

Summary Table

Leverage Point Status Priority Recommendation

Context Good/Fair/Poor High/Med/Low Specific action

... ... ... ...

Priority Actions

List top 3-5 improvements in order of impact:

  • [Highest Impact] - Specific recommendation

  • [High Impact] - Specific recommendation

  • [Medium Impact] - Specific recommendation

Detailed Findings

For each leverage point:

  • Current state

  • Specific gaps found

  • Recommended improvements

  • Example of what good looks like

Example Audit Output

Leverage Point Audit Results

Summary

  • Tests: POOR (no test files found) - HIGHEST PRIORITY
  • Standard Out: FAIR (some logging, missing error cases)
  • Architecture: GOOD (clear structure, reasonable file sizes)

Priority Actions

  1. Add test suite - enables self-correction
  2. Add error logging to API endpoints - enables visibility
  3. Create /prime command - enables quick context

Detailed Findings

[... specific recommendations ...]

Related Memory Files

  • @12-leverage-points.md - Complete framework reference

  • @agentic-kpis.md - How to measure improvement

  • @agent-perspective-checklist.md - Quick pre-task checklist

Version History

  • v1.0.0 (2025-12-26): Initial release

Last Updated

Date: 2025-12-26 Model: claude-opus-4-5-20251101

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

api-security

No summary provided by upstream source.

Repository SourceNeeds Review
Security

container-security

No summary provided by upstream source.

Repository SourceNeeds Review
Security

agentic-layer-audit

No summary provided by upstream source.

Repository SourceNeeds Review