project-documentation-workflow

When --yes or -y : Auto-confirm task decomposition, skip interactive validation, use defaults.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "project-documentation-workflow" with this command: npx skills add catlog22/claude-code-workflow/catlog22-claude-code-workflow-project-documentation-workflow

Auto Mode

When --yes or -y : Auto-confirm task decomposition, skip interactive validation, use defaults.

Project Documentation Workflow (Optimized)

Usage

$project-documentation-workflow "Document the authentication module in src/auth/" $project-documentation-workflow -c 4 "Generate full docs for the FEM solver project" $project-documentation-workflow -y "Document entire codebase with architecture and API" $project-documentation-workflow --continue "doc-auth-module-20260304"

Flags:

  • -y, --yes : Skip all confirmations (auto mode)

  • -c, --concurrency N : Max concurrent agents within each wave (default: 3)

  • --continue : Resume existing session

Output Directory: .workflow/.csv-wave/{session-id}/

Core Output: tasks.csv

  • results.csv
  • discoveries.ndjson
  • wave-summaries/
  • docs/ (完整文档集)

Overview

优化版:动态任务分解 + 拓扑排序波次计算 + 波次间综合步骤。

┌─────────────────────────────────────────────────────────────────────────┐ │ PROJECT DOCUMENTATION WORKFLOW (Dynamic & Optimized) │ ├─────────────────────────────────────────────────────────────────────────┤ │ │ │ Phase 0: Dynamic Decomposition │ │ ├─ Analyze project structure, complexity, domain │ │ ├─ Generate appropriate documentation tasks (动态数量) │ │ ├─ Compute task dependencies (deps) │ │ ├─ Compute execution waves (topological sort) │ │ └─ User validates task breakdown (skip if -y) │ │ │ │ Phase 1: Wave Execution (with Inter-Wave Synthesis) │ │ ├─ For each wave (1..N, dynamically computed): │ │ │ ├─ Load Wave Summary from previous wave │ │ │ ├─ Build wave CSV with prev_context injection │ │ │ ├─ spawn_agents_on_csv(wave CSV) │ │ │ ├─ Collect results, merge into master tasks.csv │ │ │ ├─ Generate Wave Summary (波次综合) │ │ │ └─ Check: any failed? → skip dependents │ │ └─ discoveries.ndjson shared across all waves │ │ │ │ Phase 2: Results Aggregation │ │ ├─ Export final results.csv │ │ ├─ Generate context.md with all findings │ │ ├─ Generate docs/index.md navigation │ │ └─ Display summary: completed/failed/skipped per wave │ │ │ └─────────────────────────────────────────────────────────────────────────┘

CSV Schema

tasks.csv (Master State)

id,title,description,doc_type,target_scope,doc_sections,formula_support,priority,deps,context_from,wave,status,findings,doc_path,key_discoveries,error "doc-001","项目概述","撰写项目的整体概述","overview","README.md,package.json","purpose,background,positioning,audience","false","high","","","1","pending","","","",""

Columns:

Column Type Required Description

id

string Yes Task ID (doc-NNN, auto-generated)

title

string Yes Document title

description

string Yes Detailed task description

doc_type

enum Yes `overview

target_scope

string Yes File scope (glob pattern)

doc_sections

string Yes Required sections (comma-separated)

formula_support

boolean No LaTeX formula support

priority

enum No `high

deps

string No Dependency task IDs (semicolon-separated)

context_from

string No Context source task IDs

wave

integer Computed Wave number (computed by topological sort)

status

enum Output `pending→completed

findings

string Output Key findings summary

doc_path

string Output Generated document path

key_discoveries

string Output Key discoveries (JSON)

error

string Output Error message

Implementation

Session Initialization

const getUtc8ISOString = () => new Date(Date.now() + 8 * 60 * 60 * 1000).toISOString()

const AUTO_YES = $ARGUMENTS.includes('--yes') || $ARGUMENTS.includes('-y') const continueMode = $ARGUMENTS.includes('--continue') const concurrencyMatch = $ARGUMENTS.match(/(?:--concurrency|-c)\s+(\d+)/) const maxConcurrency = concurrencyMatch ? parseInt(concurrencyMatch[1]) : 3

const requirement = $ARGUMENTS .replace(/--yes|-y|--continue|--concurrency\s+\d+|-c\s+\d+/g, '') .trim()

const slug = requirement.toLowerCase() .replace(/[^a-z0-9\u4e00-\u9fa5]+/g, '-') .substring(0, 40) const dateStr = getUtc8ISOString().substring(0, 10).replace(/-/g, '') const sessionId = doc-${slug}-${dateStr} const sessionFolder = .workflow/.csv-wave/${sessionId}

Bash(mkdir -p ${sessionFolder}/docs ${sessionFolder}/wave-summaries)

// Initialize discoveries.ndjson Write(${sessionFolder}/discoveries.ndjson, # Discovery Board - ${sessionId}\n# Format: NDJSON\n)

Phase 0: Dynamic Task Decomposition

Objective: Analyze the project and dynamically generate appropriate documentation tasks.

Step 1: Project Analysis

Bash({ command: `ccw cli -p "PURPOSE: Analyze the project and determine appropriate documentation tasks. TASK:

  1. Scan project structure to identify:

    • Project type (library/application/service/CLI/tool)
    • Primary language(s) and frameworks
    • Project scale (small/medium/large based on file count and complexity)
    • Key modules and their purposes
    • Existing documentation (README, docs/, etc.)
  2. Determine documentation needs based on project characteristics:

    • For ALL projects: overview, tech-stack, directory-structure
    • For libraries: api-reference, usage-guide, best-practices
    • For applications: system-architecture, feature-list, usage-guide
    • For numerical/scientific projects: theoretical-foundations (with formula_support=true)
    • For services: api-reference, module-interactions, deployment
    • For complex projects (>50 files): add design-patterns, data-model
    • For simple projects (<10 files): reduce to essential docs only
  3. Generate task list with:

    • Unique task IDs (doc-001, doc-002, ...)
    • Appropriate doc_type for each task
    • Target scope (glob patterns) based on actual project structure
    • Required sections for each document type
    • Dependencies (deps) between related tasks
    • Context sources (context_from) for information flow
    • Priority (high for essential docs, medium for useful, low for optional)
  4. Task dependency rules:

    • overview tasks: no deps (Wave 1)
    • architecture tasks: depend on overview tasks
    • implementation tasks: depend on architecture tasks
    • feature/api tasks: depend on implementation
    • synthesis tasks: depend on most other tasks

MODE: analysis CONTEXT: @**/* EXPECTED: JSON with:

  • project_info: {type, scale, languages, frameworks, modules[]}
  • recommended_waves: number of waves suggested
  • tasks: [{id, title, description, doc_type, target_scope, doc_sections, formula_support, priority, deps[], context_from[]}]

CONSTRAINTS:

  • Small projects: 5-8 tasks max
  • Medium projects: 10-15 tasks
  • Large projects: 15-25 tasks
  • Each doc_type should appear at most once unless justified
  • deps must form a valid DAG (no cycles)

PROJECT TO ANALYZE: ${requirement}" --tool gemini --mode analysis --rule planning-breakdown-task-steps`, run_in_background: true })

Step 2: Topological Sort (Wave Computation)

function computeWaves(tasks) { // Build adjacency list const graph = new Map() const inDegree = new Map() const taskMap = new Map()

for (const task of tasks) { taskMap.set(task.id, task) graph.set(task.id, []) inDegree.set(task.id, 0) }

// Fill edges based on deps for (const task of tasks) { const deps = task.deps.filter(d => taskMap.has(d)) for (const dep of deps) { graph.get(dep).push(task.id) inDegree.set(task.id, inDegree.get(task.id) + 1) } }

// Kahn's BFS algorithm const waves = [] let currentWave = []

// Start with tasks that have no dependencies for (const [id, degree] of inDegree) { if (degree === 0) currentWave.push(id) }

while (currentWave.length > 0) { waves.push([...currentWave]) const nextWave = []

for (const id of currentWave) {
  for (const neighbor of graph.get(id)) {
    inDegree.set(neighbor, inDegree.get(neighbor) - 1)
    if (inDegree.get(neighbor) === 0) {
      nextWave.push(neighbor)
    }
  }
}

currentWave = nextWave

}

// Assign wave numbers for (let w = 0; w < waves.length; w++) { for (const id of waves[w]) { taskMap.get(id).wave = w + 1 } }

// Check for cycles const assignedCount = tasks.filter(t => t.wave > 0).length if (assignedCount < tasks.length) { throw new Error(Circular dependency detected! Only ${assignedCount}/${tasks.length} tasks assigned.) }

return { tasks: tasks, waveCount: waves.length, waveDistribution: waves.map((w, i) => ({ wave: i + 1, tasks: w.length })) } }

Step 3: User Validation

// Parse decomposition result const analysisResult = JSON.parse(decompositionOutput) const { tasks, project_info, waveCount } = analysisResult

// Compute waves const { tasks: tasksWithWaves, waveCount: computedWaves, waveDistribution } = computeWaves(tasks)

// Display to user (skip if AUTO_YES) if (!AUTO_YES) { console.log(╔════════════════════════════════════════════════════════════════╗ ║ PROJECT ANALYSIS RESULT ║ ╠════════════════════════════════════════════════════════════════╣ ║ Type: ${project_info.type.padEnd(20)} Scale: ${project_info.scale.padEnd(10)} ║ ║ Languages: ${project_info.languages.join(', ').substring(0, 40).padEnd(40)} ║ ║ Modules: ${project_info.modules.length} identified ║ ╠════════════════════════════════════════════════════════════════╣ ║ WAVE DISTRIBUTION (${computedWaves} waves, ${tasksWithWaves.length} tasks) ║ ${waveDistribution.map(w =>║ Wave ${w.wave}: ${w.tasks} tasks${' '.repeat(50 - w.tasks.toString().length)}).join('\n')} ╚════════════════════════════════════════════════════════════════╝ )

// Show tasks by wave for (let w = 1; w <= computedWaves; w++) { const waveTasks = tasksWithWaves.filter(t => t.wave === w) console.log(\nWave ${w}:) for (const t of waveTasks) { console.log( ${t.id}: ${t.title} [${t.doc_type}]) } }

const confirm = AskUserQuestion("Proceed with this task breakdown?") if (!confirm) { console.log("Aborted. Use --continue to resume with modified tasks.") return } }

// Generate tasks.csv Write(${sessionFolder}/tasks.csv, toCsv(tasksWithWaves)) Write(${sessionFolder}/project-info.json, JSON.stringify(project_info, null, 2))

Phase 1: Wave Execution (with Inter-Wave Synthesis)

Key Optimization: Add Wave Summary generation between waves for better context propagation.

const masterCsv = Read(${sessionFolder}/tasks.csv) let tasks = parseCsv(masterCsv) const maxWave = Math.max(...tasks.map(t => t.wave))

for (let wave = 1; wave <= maxWave; wave++) { console.log(\n{'='*60}) console.log(Wave ${wave}/${maxWave}) console.log('='.repeat(60))

// 1. Load Wave Summary from previous wave const waveSummaryPath = ${sessionFolder}/wave-summaries/wave-${wave-1}-summary.md let prevWaveSummary = '' if (wave > 1 && fileExists(waveSummaryPath)) { prevWaveSummary = Read(waveSummaryPath) console.log(Loaded Wave ${wave-1} Summary (${prevWaveSummary.length} chars)) }

// 2. Filter tasks for this wave const waveTasks = tasks.filter(t => t.wave === wave && t.status === 'pending')

// 3. Check dependencies for (const task of waveTasks) { const depIds = (task.deps || '').split(';').filter(Boolean) const depStatuses = depIds.map(id => tasks.find(t => t.id === id)?.status) if (depStatuses.some(s => s === 'failed' || s === 'skipped')) { task.status = 'skipped' task.error = Dependency failed: ${depIds.filter((id, i) => ['failed','skipped'].includes(depStatuses[i])).join(', ')} } }

const pendingTasks = waveTasks.filter(t => t.status === 'pending') if (pendingTasks.length === 0) { console.log(Wave ${wave}: No pending tasks, skipping...) continue }

// 4. Build enhanced prev_context for (const task of pendingTasks) { // a. From context_from tasks const contextIds = (task.context_from || '').split(';').filter(Boolean) const prevFindings = contextIds.map(id => { const src = tasks.find(t => t.id === id) if (!src?.findings) return '' return ## [${src.id}] ${src.title}\n${src.findings} }).filter(Boolean).join('\n\n')

// b. From previous wave summary (HIGH DENSITY CONTEXT)
const waveContext = prevWaveSummary ? 
  `\n\n## Wave ${wave-1} Summary\n${prevWaveSummary}` : ''

// c. From discoveries.ndjson (relevant entries)
const discoveries = Read(`${sessionFolder}/discoveries.ndjson`)
const relevantDiscoveries = discoveries
  .split('\n')
  .filter(line => line.startsWith('{'))
  .map(line => JSON.parse(line))
  .filter(d => isRelevantDiscovery(d, task))
  .slice(0, 10) // Limit to 10 most relevant
  .map(d => `- [${d.type}] ${JSON.stringify(d.data)}`)
  .join('\n')

const discoveryContext = relevantDiscoveries ? 
  `\n\n## Relevant Discoveries\n${relevantDiscoveries}` : ''

task.prev_context = prevFindings + waveContext + discoveryContext

}

// 5. Write wave CSV Write(${sessionFolder}/wave-${wave}.csv, toCsv(pendingTasks))

// 6. Execute wave spawn_agents_on_csv({ csv_path: ${sessionFolder}/wave-${wave}.csv, id_column: "id", instruction: buildOptimizedInstruction(sessionFolder, wave), max_concurrency: maxConcurrency, max_runtime_seconds: 900, output_csv_path: ${sessionFolder}/wave-${wave}-results.csv, output_schema: { type: "object", properties: { id: { type: "string" }, status: { type: "string", enum: ["completed", "failed"] }, findings: { type: "string" }, doc_path: { type: "string" }, key_discoveries: { type: "string" }, error: { type: "string" } } } })

// 7. Merge results const results = parseCsv(Read(${sessionFolder}/wave-${wave}-results.csv)) for (const r of results) { const t = tasks.find(t => t.id === r.id) if (t) Object.assign(t, r) } Write(${sessionFolder}/tasks.csv, toCsv(tasks))

// 8. Generate Wave Summary (NEW: Inter-Wave Synthesis) const completedThisWave = results.filter(r => r.status === 'completed') if (completedThisWave.length > 0) { const waveSummary = generateWaveSummary(wave, completedThisWave, tasks) Write(${sessionFolder}/wave-summaries/wave-${wave}-summary.md, waveSummary) console.log(Generated Wave ${wave} Summary) }

// 9. Cleanup temp files Bash(rm -f ${sessionFolder}/wave-${wave}.csv ${sessionFolder}/wave-${wave}-results.csv)

// 10. Display wave summary const completed = results.filter(r => r.status === 'completed').length const failed = results.filter(r => r.status === 'failed').length console.log(Wave ${wave} Complete: ${completed} completed, ${failed} failed) }

Wave Summary Generation (Inter-Wave Synthesis)

function generateWaveSummary(waveNum, completedTasks, allTasks) { let summary = # Wave ${waveNum} Summary\n\n summary += **Completed Tasks**: ${completedTasks.length}\n\n

// Group by doc_type const byType = {} for (const task of completedTasks) { const type = task.doc_type || 'unknown' if (!byType[type]) byType[type] = [] byType[type].push(task) }

for (const [type, tasks] of Object.entries(byType)) { summary += ## ${type.toUpperCase()}\n\n for (const t of tasks) { summary += ### ${t.title}\n if (t.findings) { summary += ${t.findings.substring(0, 300)}${t.findings.length > 300 ? '...' : ''}\n\n } if (t.key_discoveries) { try { const discoveries = JSON.parse(t.key_discoveries) summary += **Key Points**:\n for (const d of discoveries.slice(0, 3)) { summary += - ${d.name || d.type}: ${d.description || JSON.stringify(d).substring(0, 100)}\n } summary += '\n' } catch (e) {} } } }

// Add cross-references for next wave const nextWaveTasks = allTasks.filter(t => t.wave === waveNum + 1) if (nextWaveTasks.length > 0) { summary += ## Context for Wave ${waveNum + 1}\n\n summary += Next wave will focus on: ${nextWaveTasks.map(t => t.title).join(', ')}\n }

return summary }

function isRelevantDiscovery(discovery, task) { // Check if discovery is relevant to the task const taskScope = task.target_scope || '' const taskType = task.doc_type || ''

// Always include architecture discoveries for architecture tasks if (taskType === 'architecture' && discovery.type.includes('component')) return true if (taskType === 'implementation' && discovery.type.includes('algorithm')) return true if (taskType === 'api' && discovery.type.includes('api')) return true

// Check file relevance if (discovery.data?.file) { return taskScope.includes(discovery.data.file.split('/')[0]) }

return false }

Optimized Instruction Template

function buildOptimizedInstruction(sessionFolder, wave) { return `## DOCUMENTATION TASK — Wave ${wave}

⚠️ MANDATORY FIRST STEPS (DO NOT SKIP)

  1. CHECK DISCOVERIES FIRST (避免重复工作): ```bash

    Search for existing discoveries about your topic

    grep -i "{doc_type}" ${sessionFolder}/discoveries.ndjson grep -i "{target_keywords}" ${sessionFolder}/discoveries.ndjson ```

  2. Read Wave Summary (高密度上下文):

    • Read: ${sessionFolder}/wave-summaries/wave-${wave-1}-summary.md (if exists)
  3. Read prev_context (provided below)


Your Task

Task ID: {id} Title: {title} Document Type: {doc_type} Target Scope: {target_scope} Required Sections: {doc_sections} LaTeX Support: {formula_support} Priority: {priority}

Task Description

{description}

Previous Context (USE THIS!)

{prev_context}


Execution Protocol

Step 1: Discovery Check (MANDATORY)

Before reading any source files:

  • Search discoveries.ndjson for existing findings
  • Note any pre-discovered components, patterns, algorithms
  • Avoid re-documenting what's already found

Step 2: Scope Analysis

  • Read files matching `{target_scope}`
  • Identify key structures, functions, classes
  • Extract relevant code patterns

Step 3: Context Integration

  • Build on findings from prev_context
  • Reference Wave Summary insights
  • Connect to discoveries from other agents

Step 4: Document Generation

Output Path: Determine based on doc_type:

  • `overview` → `docs/01-overview/`
  • `architecture` → `docs/02-architecture/`
  • `implementation` → `docs/03-implementation/`
  • `feature` → `docs/04-features/`
  • `api` → `docs/04-features/`
  • `usage` → `docs/04-features/`
  • `synthesis` → `docs/05-synthesis/`

Document Structure: ```markdown

{Title}

Overview

[Brief introduction]

{Required Section 1}

[Content with code examples]

{Required Section 2}

[Content with diagrams if applicable]

...

Code Examples

```{language} // file:line references ```

Cross-References

Summary

[Key takeaways] ```

Step 5: Share Discoveries (MANDATORY)

Append to discovery board: ```bash echo '{"ts":"${getUtc8ISOString()}","worker":"{id}","type":"<TYPE>","data":{...}}' >> ${sessionFolder}/discoveries.ndjson ```

Discovery Types:

  • `component_found`: {name, type, file, purpose}
  • `pattern_found`: {pattern_name, location, description}
  • `algorithm_found`: {name, file, complexity, purpose}
  • `formula_found`: {name, latex, file, context}
  • `feature_found`: {name, entry_point, description}
  • `api_found`: {endpoint, file, parameters, returns}
  • `config_found`: {name, file, type, default_value}

Step 6: Report

```json { "id": "{id}", "status": "completed", "findings": "Key discoveries (max 500 chars, structured for context propagation)", "doc_path": "docs/XX-category/filename.md", "key_discoveries": "[{"name":"...","type":"...","description":"...","file":"..."}]", "error": "" } ```


Quality Requirements

RequirementCriteria
Section CoverageALL sections in doc_sections present
Code ReferencesInclude file:line for code
Discovery SharingAt least 2 discoveries shared
Context UsageReference prev_context findings
Cross-ReferencesLink to related docs
`
}

Phase 2: Results Aggregation

// 1. Generate docs/index.md const tasks = parseCsv(Read(${sessionFolder}/tasks.csv)) const completed = tasks.filter(t => t.status === 'completed')

// Group by doc_type for navigation const byType = {} for (const t of completed) { const type = t.doc_type || 'other' if (!byType[type]) byType[type] = [] byType[type].push(t) }

let index = # Project Documentation Index\n\n index += **Generated**: ${getUtc8ISOString().substring(0, 10)}\n index += **Total Documents**: ${completed.length}\n\n

const typeLabels = { overview: '📋 概览', architecture: '🏗️ 架构', implementation: '⚙️ 实现', theory: '📐 理论', feature: '✨ 功能', api: '🔌 API', usage: '📖 使用', synthesis: '💡 综合' }

for (const [type, typeTasks] of Object.entries(byType)) { const label = typeLabels[type] || type index += ## ${label}\n\n for (const t of typeTasks) { index += - [${t.title}](${t.doc_path})\n } index += \n }

// Add wave summaries reference index += ## 📊 Execution Reports\n\n index += - [Wave Summaries](wave-summaries/)\n index += - [Full Context](../context.md)\n

Write(${sessionFolder}/docs/index.md, index)

// 2. Export results.csv Bash(cp ${sessionFolder}/tasks.csv ${sessionFolder}/results.csv)

// 3. Generate context.md const projectInfo = JSON.parse(Read(${sessionFolder}/project-info.json)) let contextMd = # Documentation Report\n\n contextMd += **Session**: ${sessionId}\n contextMd += **Date**: ${getUtc8ISOString().substring(0, 10)}\n\n

contextMd += ## Project Info\n contextMd += - **Type**: ${projectInfo.type}\n contextMd += - **Scale**: ${projectInfo.scale}\n contextMd += - **Languages**: ${projectInfo.languages?.join(', ') || 'N/A'}\n\n

const statusCounts = { completed: tasks.filter(t => t.status === 'completed').length, failed: tasks.filter(t => t.status === 'failed').length, skipped: tasks.filter(t => t.status === 'skipped').length } contextMd += ## Summary\n contextMd += | Status | Count |\n contextMd += |--------|-------|\n contextMd += | ✅ Completed | ${statusCounts.completed} |\n contextMd += | ❌ Failed | ${statusCounts.failed} |\n contextMd += | ⏭️ Skipped | ${statusCounts.skipped} |\n\n

// Per-wave summary const maxWave = Math.max(...tasks.map(t => t.wave)) contextMd += ## Wave Execution\n\n for (let w = 1; w <= maxWave; w++) { const waveTasks = tasks.filter(t => t.wave === w) contextMd += ### Wave ${w}\n\n for (const t of waveTasks) { const icon = t.status === 'completed' ? '✅' : t.status === 'failed' ? '❌' : '⏭️' contextMd += ${icon} **${t.title}** [${t.doc_type}]\n if (t.findings) { contextMd += ${t.findings.substring(0, 200)}${t.findings.length > 200 ? '...' : ''}\n } if (t.doc_path) { contextMd += → [${t.doc_path}](${t.doc_path})\n } contextMd += \n } }

Write(${sessionFolder}/context.md, contextMd)

// 4. Display final summary console.log(╔════════════════════════════════════════════════════════════════╗ ║ DOCUMENTATION COMPLETE ║ ╠════════════════════════════════════════════════════════════════╣ ║ ✅ Completed: ${statusCounts.completed.toString().padStart(2)} tasks ║ ║ ❌ Failed: ${statusCounts.failed.toString().padStart(2)} tasks ║ ║ ⏭️ Skipped: ${statusCounts.skipped.toString().padStart(2)} tasks ║ ╠════════════════════════════════════════════════════════════════╣ ║ Output: ${sessionFolder.padEnd(50)} ║ ╚════════════════════════════════════════════════════════════════╝)

Optimized Output Structure

.workflow/.csv-wave/doc-{slug}-{date}/ ├── project-info.json # 项目分析结果 ├── tasks.csv # Master CSV (动态生成的任务) ├── results.csv # 最终结果 ├── discoveries.ndjson # 发现板 ├── context.md # 执行报告 │ ├── wave-summaries/ # NEW: 波次摘要 │ ├── wave-1-summary.md │ ├── wave-2-summary.md │ └── ... │ └── docs/ ├── index.md # 文档导航 ├── 01-overview/ ├── 02-architecture/ ├── 03-implementation/ ├── 04-features/ └── 05-synthesis/

Optimization Summary

优化点 原版 优化版

任务数量 固定17任务 动态生成 (5-25基于项目规模)

波次计算 硬编码5波 拓扑排序动态计算

上下文传播 仅 prev_context prev_context + Wave Summary + Discoveries

发现利用 依赖自觉 强制第一步检查

文档密度 原始 findings 结构化 Wave Summary

Core Rules

  • Dynamic First: 任务列表动态生成,不预设

  • Wave Order is Sacred: 波次由拓扑排序决定

  • Discovery Check Mandatory: 必须先检查发现板

  • Wave Summary: 每波次结束生成摘要

  • Context Compound: 上下文累积传播

  • Quality Gates: 每文档必须覆盖所有 doc_sections

  • DO NOT STOP: 持续执行直到所有波次完成

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

skill-generator

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

review-code

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

ccw-help

No summary provided by upstream source.

Repository SourceNeeds Review