pov-expansion

POV-Expansion: Cross-Domain Perspective Generation

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "pov-expansion" with this command: npx skills add dangeles/claude/dangeles-claude-pov-expansion

POV-Expansion: Cross-Domain Perspective Generation

Generate systematic cross-domain perspectives through boundary spanning research frameworks. Uses structure-mapping theory, convergence tracking, and transfer evaluation to identify analogous problems and transferable solutions.

When to Use

Use this skill when:

  • Stuck on a problem - Need fresh perspectives from outside your domain

  • Evaluating analogies - Want to assess if approach from field X applies to field Y

  • Seeking breakthrough insights - Ready for 2-3 hour deep exploration

  • Have structured problem - Can articulate problem clearly enough for abstraction

When NOT to Use

Do NOT use when:

  • Exploratory browsing - No specific problem statement (use /research-pipeline instead)

  • Time-constrained - Need answer in <1 hour (this workflow takes 2-3 hours minimum)

  • Well-understood problem - Already know the solution domain (direct research faster)

  • Vague problem - Cannot articulate problem clearly (refine first)

Architecture

Three-tier architecture:

  • Orchestrator: pov-expansion-pm (coordinates workflow, handles exceptions)

  • Specialist Agents: 5 new agents for abstraction, perspective generation, synthesis, transfer evaluation

  • Supporting Skills: Reuses requirements-analyst, fact-checker, devils-advocate, editor, strategist

Pre-flight Validation

Before invoking /pov-expansion , verify:

  • Problem statement is specific (not "improve product" but "reduce onboarding time")

  • Expected runtime: 2-3 hours (workflow cannot be rushed)

  • Dependencies available: requirements-analyst, fact-checker, devils-advocate, editor, strategist skills exist

The 12-Stage Pipeline

Stage 0: Archival Guidelines Review

Owner: pov-expansion-pm (automatic) Duration: 2-5 min Checkpoint: Never (always runs automatically) Session Setup: Creates /tmp/pov-session-{YYYYMMDD-HHMMSS}-{PID}/

Initialize workflow session and extract archival guidelines from project CLAUDE.md.

Process:

  • Create session directory: /tmp/pov-session-$(date +%Y%m%d-%H%M%S)-$$/

  • Read project CLAUDE.md (if exists in working directory or parent)

  • Extract archival guidelines:

  • Repository organization (directory structure for cross-domain analyses)

  • Naming conventions (analysis- prefix for POV expansion outputs)

  • Git rules (commit after edits, no version-numbered files)

  • Document structure requirements (Executive Summary, Key Parameters Table)

  • Citation format (Nature-style inline)

  • Write archival summary to session directory: archival-guidelines-summary.md

  • Store session path in workflow state for downstream agents

Output:

session_setup: session_dir: "/tmp/pov-session-{timestamp}-{pid}/" archival_summary_path: "{session_dir}/archival-guidelines-summary.md" guidelines_found: boolean guidelines_source: string # Path to CLAUDE.md or "defaults"

Quality Gate: Session directory created, archival summary written.

Failure Handling:

  • CLAUDE.md not found: Use sensible defaults, log warning

  • Session directory creation fails: ABORT (cannot proceed without session isolation)

Stage 1: Problem Refinement

Owner: requirements-analyst (reused) Duration: 15-30 min Receives: Session directory path from Stage 0 Quality Gate: Specific problem statement with measurable criteria

Stage 2: Problem Abstraction

Owner: pov-abstractor-classifier Duration: 15-30 min Quality Gate: Rasmussen hierarchy populated (5/5 levels)

Stage 3: Domain Classification

Owner: pov-abstractor-classifier Duration: 20-45 min (includes user confirmation) Quality Gate: 4 domains proposed (1+ Near, 1+ Mid, 1+ Far), user confirmed

Stage 4: Parallel Perspective Generation

Owner: pov-perspective-analyst (x4 parallel) Duration: 30-60 min (parallel) Quality Gate: 3+ perspectives complete, each with structural analogies

Stage 5: Parallel Verification

Owner: fact-checker (x2 parallel, reused) Duration: 15-30 min (parallel) Quality Gate: No factual errors in central claims

Stage 6: Convergence Analysis

Owner: pov-synthesizer Duration: 20-40 min Quality Gate: Convergence floor met (1+ theme in 2+ reports)

Stage 7: Transfer Evaluation

Owner: pov-transfer-evaluator Duration: 20-40 min Quality Gate: Transfer feasibility scored, Far field warnings added

Stage 8: Master Synthesis

Owner: pov-synthesizer Duration: 45-90 min Quality Gate: All perspectives integrated, convergent insights highlighted

Stage 9: Strategic Assessment

Owner: strategist (reused skill) Duration: 20-40 min Quality Gate: 3+ actionable recommendations with implementation paths

Stage 10: Adversarial Review

Owner: devils-advocate (reused) Duration: 20-40 min Quality Gate: Critical issues identified, mitigations proposed

Stage 11: Editorial Polish

Owner: editor (reused) Duration: 15-30 min Quality Gate: Consistent voice, no substantive errors

Timeout Configuration

Stage Timeout Exceeded Action

0 (Archival) 5 min ABORT - cannot proceed without session

1 (Refinement) 30 min Escalate to user

2 (Abstraction) 30 min Escalate to user

3 (Classification) 45 min Proceed with HIGH confidence domains only

4 (Perspectives) 60 min total Proceed with available (minimum 2 required)

4 (per agent) 20 min Retry once, then skip

5 (Verification) 30 min Skip verification (flag risk)

6 (Convergence) 40 min Proceed with portfolio framing

7 (Transfer) 40 min Lower confidence, proceed

8 (Synthesis) 90 min Escalate to user

9 (Strategic) 40 min Skip (optional)

10 (Review) 40 min Deliver without review (flag risk)

11 (Editorial) 30 min Deliver as-is, cleanup session

Global Workflow Timeout: 8 hours On Global Timeout: Save state, escalate to user

Session Cleanup:

  • On successful completion (Stage 11 complete): Delete session directory

  • On failure/abort: Retain session directory for debugging (log path to user)

Quality Gate Specifications

Gate 0: Session Initialization (Stage 0)

Check Threshold Automated?

Session directory created exists and writable Yes

Archival summary written file exists Yes

Session path stored in workflow state path present Yes

Pass Threshold: 3/3 checks pass On Failure: ABORT - cannot proceed without session isolation

Gate 1: Problem Abstraction (Stage 2)

Check Threshold Automated?

Rasmussen levels populated 5/5 levels Yes

Structural elements identified

= 3 elements Yes

Constraints defined

= 1 constraint Yes

Success metrics measurable

= 1 metric Manual

Pass Threshold: 4/4 checks pass On Failure: Return to Stage 1 with feedback (max 2 iterations)

Gate 2: Domain Classification (Stage 3)

Check Threshold Automated?

Domains proposed 4 domains Yes

Near field present

= 1 Yes

Far field present

= 1 Yes

User confirmation Required No

Pass Threshold: All checks pass On Failure: Re-run with user guidance

Gate 3: Perspective Generation (Stage 4)

Check Threshold Automated?

Perspectives complete

= 3 of 4 Yes

Structural analogies per perspective

= 2 Yes

Domain-specific examples

= 2 per perspective Yes

No placeholder text 0 instances Yes

Pass Threshold: All checks pass Marginal Pass: 3/4 checks (flag in synthesis) On Failure: Retry failed perspective or proceed with reduced coverage

Gate 4: Transfer Feasibility (Stage 7)

Check Threshold Automated?

All perspectives evaluated 100% Yes

Structural similarity scored All have scores Yes

At least one high-transfer analogy

= 1 with score > 0.6 Yes

Far field warnings added All Far field have warnings Yes

Pass Threshold: 4/4 checks pass Marginal Pass: 3/4 (proceed with caution flag) On Failure: Return to Stage 3 with broadened domains

Gate 5: Synthesis Coherence (Stage 8)

Check Threshold Automated?

All perspectives referenced 100% Yes

Convergent insights highlighted

= 1 Yes

Contradictions resolved 0 unresolved Manual

Executive summary present Yes Yes

Pass Threshold: All checks pass On Failure: Return to synthesis with feedback

Gate 6: Strategic Actionability (Stage 9)

Check Threshold Automated?

Actionable recommendations

= 3 Yes

Implementation paths Each has path Yes

Timeline/effort estimates Present Yes

Pass Threshold: All checks pass On Failure: Add missing elements or skip stage

Session Directory Structure

Location: /tmp/pov-session-{YYYYMMDD-HHMMSS}-{PID}/

/tmp/pov-session-{YYYYMMDD-HHMMSS}-{PID}/ ├── archival-guidelines-summary.md # Stage 0 output ├── workflow-state.yaml # Workflow state for resume ├── progress.md # Progress tracking ├── stage-1-refinement.md # Stage 1 output ├── stage-2-abstraction.yaml # Stage 2 output (structured) ├── stage-3-domains.yaml # Stage 3 output (structured) ├── stage-4-perspectives/ # Parallel outputs │ ├── near-field-1.md │ ├── near-field-2.md │ ├── mid-field.md │ └── far-field.md ├── stage-5-verification/ # Parallel outputs │ ├── verification-1.md │ └── verification-2.md ├── stage-6-convergence.md ├── stage-7-transfer.md ├── stage-8-synthesis.md ├── stage-9-strategic.md ├── stage-10-review.md ├── stage-11-final.md # Final deliverable └── failure-log.md # Failure tracking

Workflow State Schema

workflow_state: workflow_id: string problem_statement: string started_at: ISO8601 current_stage: integer # 0-11 stages_completed: [integer]

session: # Added for session management session_dir: string # /tmp/pov-session-{...}/ archival_guidelines_path: string # Path to archival-guidelines-summary.md guidelines_found: boolean # Whether CLAUDE.md was found guidelines_source: string # Path to CLAUDE.md or "defaults" cleanup_on_complete: boolean # Default true

artifacts: stage_0: path | null # archival-guidelines-summary.md stage_1: path | null stage_2: path | null # ... etc stage_11: path | null

status: enum[in_progress, paused, completed, failed] last_checkpoint: ISO8601 error_log: [string]

Session Handling on Resume:

  • If session directory exists: Reuse existing session

  • If session directory missing: Re-run Stage 0 to recreate (non-destructive)

Error Handling Protocols

Catastrophic Failure Protocol (Stage 4)

When ALL perspective agents fail:

  • Detection: 0 valid outputs from 4 parallel agents

  • Check for partial outputs: Any content > 500 chars is "partial"

  • Present user options:

PERSPECTIVE GENERATION FAILED: All 4 agents failed

Failures:

  • Near Field 1: [status]
  • Near Field 2: [status]
  • Mid Field: [status]
  • Far Field: [status]

Options: (A) Retry SEQUENTIALLY (slower but more stable) (B) Retry with 2 agents only (Near + Mid fields) (C) Review partial outputs if available (D) Abort workflow, preserve problem abstraction

Convergence Floor Protocol

Minimum Convergence Threshold:

  • 1+ themes appearing in 2+ reports, OR

  • 2+ strategies appearing in 2+ reports, OR

  • 1+ structural analogies across 2+ domains

On Zero Convergence:

  • Re-analyze with relaxed matching (related concepts, not exact)

  • Look for "anti-convergence" (all agree something doesn't work)

  • If still zero: Offer portfolio framing (present as independent options, not validated patterns)

User Response Timeout Protocol

Gate Timeout Default Action

Domain Classification 10 min Proceed with HIGH confidence domains only

Transfer Feasibility 5 min Proceed with portfolio view

Strategic Priority 5 min Present all options without prioritization

Parallel Execution

Stage 4: Perspective Generation (Fork-Join)

parallel_execution: stage: 4 pattern: fork_join max_agents: 4

agent_assignment: agent_1: "Near Field 1" agent_2: "Near Field 2" agent_3: "Mid Field" agent_4: "Far Field"

output_paths: agent_1: "/tmp/pov-session-{id}/stage-4-perspectives/near-field-1.md" agent_2: "/tmp/pov-session-{id}/stage-4-perspectives/near-field-2.md" agent_3: "/tmp/pov-session-{id}/stage-4-perspectives/mid-field.md" agent_4: "/tmp/pov-session-{id}/stage-4-perspectives/far-field.md"

completion: minimum_required: 2 proceed_on_timeout: true log_missing: true

Stage 5: Verification (Parallel)

parallel_execution: stage: 5 pattern: parallel max_agents: 2

assignment: agent_1: "Near + Mid perspectives" agent_2: "Far perspective + cross-check"

Success Criteria

  • Diversity: Perspectives span Near, Mid, and Far fields (4 domains minimum)

  • Depth: Structural analogies via structure-mapping principles (>=2 per perspective)

  • Convergence: Cross-perspective insights tracked with measurable algorithm

  • Transferability: Rigorous transfer evaluation with Far field warnings

  • Actionability: Strategic assessment with >=3 recommendations and implementation paths

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

bioinformatician

No summary provided by upstream source.

Repository SourceNeeds Review
General

statistician

No summary provided by upstream source.

Repository SourceNeeds Review
General

procurement

No summary provided by upstream source.

Repository SourceNeeds Review
General

requirements-analyst

No summary provided by upstream source.

Repository SourceNeeds Review