Plan Review
Comprehensive review of implementation plans using parallel specialized agents.
Usage
/review-plan # Review plan from current context /review-plan path/to/plan.md # Review specific plan file
Workflow
Step 1: Extract Plan and Check Internal Docs
Get the plan to review:
-
If path provided, read the file
-
Otherwise, use the plan from current conversation context
-
Summarize: problem statement, proposed solution, key implementation steps
Check internal documentation: Use Grep to search for relevant keywords in docs/ and *.md files. Look for documented patterns, architectural guidelines, or gotchas related to the plan's area.
Step 2: Determine Review Scope
Based on plan complexity, decide:
-
Simple (single file, minor change): Skip research agent, 2 alternatives
-
Medium (few files, new feature): All agents, 3 alternatives
-
Complex (architectural, multi-system): All agents + research, 4 alternatives
Do NOT shortcut this workflow:
-
"I already know the issues" -- External perspectives find blind spots you can't see
-
"This will take too long" -- Parallel agents run simultaneously, the time cost is minimal
Step 3: Spawn Review Agents in Parallel
CRITICAL: Launch agents in a SINGLE message with multiple tool calls. Do NOT invoke one at a time. Do NOT stop after the first agent.
Agent Purpose Tool
External Opinions Get Gemini + Codex input Skill: second-opinion
Alternatives Propose 2-4 other solutions Task: general-purpose
Robustness Check for fragile patterns Task: general-purpose
Adversarial Maximally critical review Task: general-purpose
Research Relevant practices online Skill: research-online
See references/agent-prompts.md for full prompt templates for each agent.
Step 4: Synthesize Findings
Collect all agent results and synthesize:
Plan Review: {plan_name}
External Opinions
Gemini: {summary} Codex: {summary} Consensus: {where they agree} Divergence: {where they disagree}
Alternative Approaches
| Approach | Key Advantage | Key Disadvantage |
|---|---|---|
| Current plan | {pro} | {con} |
| Alt 1: {name} | {pro} | {con} |
| Alt 2: {name} | {pro} | {con} |
Recommendation: {stick with plan / consider alternative X / hybrid}
Robustness Issues
Critical (must fix):
- {issue}: {fix}
Warnings:
- {issue}: {fix}
Adversarial Findings
Valid concerns:
- {concern}: {how to address}
Dismissed concerns:
- {concern}: {why it's not a real issue}
Research Insights
(if applicable)
- {relevant finding}
Revised Plan Recommendations
{specific improvements to make based on all feedback}
Changes to Make
- {change 1}
- {change 2}
Questions to Resolve
- {unresolved question}
Step 5: Update Plan
If significant issues found, offer to revise the plan incorporating the feedback.
Examples
Review a refactor plan -- agents find a robustness issue:
/review-plan
Spawns parallel review agents against the current plan. The robustness agent flags that the migration has no rollback path if it fails midway, and the adversarial agent identifies a race condition under concurrent writes. The synthesis recommends adding a rollback step and a distributed lock.
Review an auth plan with research agent:
/review-plan docs/plans/auth-redesign.md
Reviews the auth redesign plan with all agents including the research agent, which finds that the proposed token rotation strategy has a known edge case documented in the OAuth 2.1 spec. The synthesis recommends adjusting the refresh window based on the research findings.
Troubleshooting
Review agents disagree on approach
Solution: Focus on the points of consensus first, then evaluate the disagreements by weighing each agent's reasoning against your project constraints. Use the adversarial agent's concerns as a tiebreaker -- if it flags real risk in one approach, prefer the safer alternative.
Plan is too vague for meaningful review
Solution: Add concrete details before running the review: specify which files change, what data flows through the system, and what the failure modes are. Agents produce generic feedback when the plan lacks implementation-level specifics.
Notes
-
Use the Skill tool for second-opinion and research-online
-
do not write slash commands directly
-
External opinions provide model diversity (Gemini + Codex)
-
The adversarial agent should be harsh - that's its job
-
Robustness review catches patterns that "work in testing, fail in prod" - see references/robustness-patterns.md for examples
-
Research agent finds relevant practices and known issues online
-
Always synthesize all agent results into actionable improvements