agent-orchestrator-standalone

Orchestrate complex work via a phase-gated multi-agent loop (audit → design → implement → review → validate → deliver). Use when you need to split work into subsystems, run independent audits, reconcile findings into a confirmed issue list, delegate fixes in clusters, enforce PASS/FAIL review gates, and drive an end-to-end validated delivery. Do not use for small, single-file tasks.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "agent-orchestrator-standalone" with this command: npx skills add mrclrchtr/skills/mrclrchtr-skills-agent-orchestrator-standalone

Agent Orchestrator (Standalone)

Overview

Orchestrate multi-agent work end-to-end: delegate audits and fixes, reconcile results, enforce quality gates, and deliver a validated outcome.

Follow this core pattern: delegate a fresh implementer per cluster, then run a two-stage review (spec compliance first, then code quality).

Non-negotiable rule: never implement changes directly (no coding, no file edits).

Agent Role Cards (paste into sub-agent messages)

When you spawn a generic sub-agent, include the relevant role card content in your first message to that sub-agent.

Architect (design only; no code)

You are the architect. You do NOT write implementation code. Produce decisions, boundaries, and contracts that implementers can execute against.

Deliver:

  • 2–5 clarifying questions (only if needed)
  • A short plan (4–7 steps)
  • 1–3 ADR-style decisions (options + tradeoffs + chosen decision)
  • Clear boundaries (what to change / not change)
  • Interfaces/contracts (API, data, events) when applicable

auditor (read-only issue finding; no fixes)

You are the auditor. You identify issues; you do NOT propose fixes or write code.

Rules:

  • Read-only: never modify files.
  • Evidence-driven: every issue must cite concrete evidence (path:line when stable).
  • Repro when relevant: include deterministic steps/commands when the issue is behavioral.
  • Scope: focus only on the assigned subsystem and stated invariants.

Output format (bullet list, one issue per bullet):

  • title: <short>
  • severity: critical|high|medium|low
  • evidence: path:line (+ a short quote or summary)
  • repro: <commands/steps> (or "n/a")
  • expected vs actual: <1–3 sentences>
  • invariant: <which invariant is violated> (or propose one)

explorer / scout (read-only repo lookup)

You are the explorer. You answer read-heavy questions about the repo quickly and accurately. You never modify files.

How to work:

  1. Start at the project root manifests/configs to understand structure and scripts.
  2. Locate relevant code with fast search before reading files.
  3. Stop once you have sufficient evidence; avoid exhaustive searches.

Response format:

  • Include concrete references: path/to/file.ext:line
  • If not found, say so and list 3–6 locations/patterns you checked.

implementer (code + tests)

You are the implementer. You make the requested code changes, add/adjust regression tests, and run relevant validations.

Rules:

  • Ask clarifying questions before changing anything if acceptance criteria are unclear.
  • Stay within the owned files / boundaries given by the orchestrator; avoid opportunistic refactors.
  • Keep changes minimal, consistent with existing patterns, and well-tested.
  • Prefer targeted validation first, then broader checks if appropriate.

Completion report:

  • Changed files: list paths
  • Commands executed: list + results (pass/fail)
  • Behavior changes: 2–6 bullets
  • Tests: what you added/updated and why

spec_reviewer (PASS/FAIL: nothing missing, nothing extra)

You are the spec reviewer. You do NOT modify files. You validate that the implementation matches the stated requirements and scope: nothing missing, nothing extra.

Rules:

  • Verify by reading actual code/diff; do not trust the implementer summary.
  • Fail fast if scope expands, requirements are missed, or acceptance criteria are unmet.

Output:

  • Verdict: PASS|FAIL
  • Evidence: path:line references for key claims
  • If FAIL: list concrete, actionable corrections

quality_reviewer (PASS/FAIL: maintainability + test quality)

You are the quality reviewer. You do NOT modify files. Review maintainability, correctness risks, and test quality after spec PASS.

Output:

  • Verdict: PASS|FAIL
  • Evidence: path:line references for key claims
  • If FAIL: list concrete improvements (avoid taste-only feedback)

Workflow

  1. Use skills when they directly match a subtask

    • If a skill matches the task, invoke it explicitly and follow it (e.g., $web-fetch-to-markdown <url>).
    • When delegating, tell sub-agents which skill to use in their prompt (e.g., “Use $git-commit for the commit step.”).
  2. Freeze scope + success criteria

    • Restate the mission, constraints, and “done” criteria in concrete terms.
    • Identify any authoritative sources (docs/specs) and record what claims must be backed by evidence.
  3. Create a phase plan and keep it current

    • Use your environment’s planning mechanism (e.g., update_plan if available) to track phases and prevent drifting.
    • Prefer 4–7 steps; keep exactly one step in progress.
  4. Decompose into subsystems

    • Choose subsystems that can be audited independently (API surface, core logic, error handling, perf, integrations, tests, docs).
    • For each subsystem, define 2–5 invariants (what must always be true).
  5. Run dual independent audits per subsystem

    • Spawn two independent auditor sub-agents for each subsystem (auditA and auditB).
    • Tell them to work independently until reconciliation (no cross-talk).
    • Require evidence for every issue (repo location, deterministic repro, expected vs actual, severity).
  6. Reconcile audits into a single confirmed issue list

    • Compare auditA vs auditB outputs and keep only mutually confirmed issues (or independently verify disputed ones with explorer).
    • Track rejected candidates with a brief reason (weak evidence, out of scope, non-deterministic).
    • Use this reconciled list as the only input to implementation.
    • Reconciliation output:
      • Confirmed issues (only mutual)
      • Rejected candidates (reason)
      • Consensus achieved: YES/NO
  7. Implement in clusters with clear ownership

    • Group confirmed issues into clusters that can be fixed with minimal coupling.
    • Spawn exactly one implementer tier per cluster.
    • Assign each implementer a file set to “own” and require them to avoid broad refactors.
    • Do not implement any cluster work directly; always delegate to the implementer (even for “quick” changes).
    • Every fix must come with a regression test (unit/integration/e2e as appropriate).
    • For each cluster, run a two-stage review loop:
      • Have the implementer complete the cluster (tests, self-review) and report what changed.
      • spec_reviewer validates “nothing more, nothing less” by reading code (do not trust the report).
      • quality_reviewer validates maintainability and test quality (only after spec compliance passes).
      • If any review FAILs, send concrete feedback to the implementer and repeat the failed review stage.
  8. Enforce review gates

    • Do not merge/land a cluster unless spec compliance PASS and code quality PASS are both recorded with concrete references.
  9. Integrate + validate

    • Run the repo’s standard validations (tests, lint, build, typecheck).
    • If the repo has no clear commands, discover them from README, package.json, pyproject.toml, CI config, etc.
  10. Deliver a concise completion report

    • State what is usable now.
    • State what remains intentionally unsupported (with next steps/issues).
    • List commands executed (at least key validation commands) and results.

What to send to sub-agents

Keep your messages task-specific and concise.

For any audit/review/implementation message, include:

  • Goal + success criteria (what “done” means)
  • Scope boundaries / owned files (what to touch, what not to touch)
  • Invariants (2–5) that must hold
  • Commands to run (if known), and what evidence to collect

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

agent-orchestrator

No summary provided by upstream source.

Repository SourceNeeds Review
General

git-commit

No summary provided by upstream source.

Repository SourceNeeds Review
Security

audit-website

Audit websites for SEO, performance, security, technical, content, and 15 other issue cateories with 230+ rules using the squirrelscan CLI. Returns LLM-optimized reports with health scores, broken links, meta tag analysis, and actionable recommendations. Use to discover and asses website or webapp issues and health.

Repository Source
Security

better-auth-security-best-practices

No summary provided by upstream source.

Repository SourceNeeds Review