Rapid Convergence

Achieve 3-4 iteration methodology convergence (vs standard 5-7) when clear baseline metrics exist, domain scope is focused, and direct validation is possible. Use when you have V_meta baseline ≥0.40, quantifiable success criteria, retrospective validation data, and generic agents are sufficient. Enables 40-60% time reduction (10-15 hours vs 20-30 hours) without sacrificing quality. Prediction model helps estimate iteration count during experiment planning. Validated in error recovery (3 iterations, 10 hours, V_instance=0.83, V_meta=0.85).

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "Rapid Convergence" with this command: npx skills add zpankz/mcp-skillset/zpankz-mcp-skillset-rapid-convergence

Rapid Convergence

Achieve methodology convergence in 3-4 iterations through structural optimization, not rushing.

Rapid convergence is not about moving fast - it's about recognizing when structural factors naturally enable faster progress without sacrificing quality.


When to Use This Skill

Use this skill when:

  • 🎯 Planning new experiment: Want to estimate iteration count and timeline
  • 📊 Clear baseline exists: Can quantify current state with V_meta(s₀) ≥ 0.40
  • 🔍 Focused domain: Can describe scope in <3 sentences without ambiguity
  • Direct validation: Can validate with historical data or single context
  • Time constraints: Need methodology in 10-15 hours vs 20-30 hours
  • 🧩 Generic agents sufficient: No complex specialization needed

Don't use when:

  • ❌ Exploratory research (no established metrics)
  • ❌ Multi-context validation required (cross-language, cross-domain testing)
  • ❌ Complex specialization needed (>10x speedup from specialists)
  • ❌ Incremental pattern discovery (patterns emerge gradually, not upfront)

Quick Start (5 minutes)

Rapid Convergence Self-Assessment

Answer these 5 questions:

  1. Baseline metrics exist: Can you quantify current state objectively? (YES/NO)
  2. Domain is focused: Can you describe scope in <3 sentences? (YES/NO)
  3. Validation is direct: Can you validate without multi-context deployment? (YES/NO)
  4. Prior art exists: Are there established practices to reference? (YES/NO)
  5. Success criteria clear: Do you know what "done" looks like? (YES/NO)

Scoring:

  • 4-5 YES: ⚡ Rapid convergence (3-4 iterations) likely
  • 2-3 YES: 📊 Standard convergence (5-7 iterations) expected
  • 0-1 YES: 🔬 Exploratory (6-10 iterations), establish baseline first

Five Rapid Convergence Criteria

Criterion 1: Clear Baseline Metrics (CRITICAL)

Indicator: V_meta(s₀) ≥ 0.40

What it means:

  • Domain has established metrics (error rate, test coverage, build time)
  • Baseline can be measured objectively in iteration 0
  • Success criteria can be quantified before starting

Example (Bootstrap-003):

✅ Clear baseline:
- 1,336 errors quantified via MCP queries
- 5.78% error rate calculated
- Clear MTTD/MTTR targets
- Result: V_meta(s₀) = 0.48

Outcome: 3 iterations, 10 hours

Counter-example (Bootstrap-002):

❌ No baseline:
- No existing test coverage data
- Had to establish metrics first
- Fuzzy success criteria initially
- Result: V_meta(s₀) = 0.04

Outcome: 6 iterations, 25.5 hours

Impact: High V_meta baseline means:

  • Fewer iterations to reach 0.80 threshold (+0.40 vs +0.76)
  • Clearer iteration objectives (gaps are obvious)
  • Faster validation (metrics already exist)

See reference/baseline-metrics.md for achieving V_meta ≥ 0.40.

Criterion 2: Focused Domain Scope (IMPORTANT)

Indicator: Domain described in <3 sentences without ambiguity

What it means:

  • Single cross-cutting concern
  • Clear boundaries (what's in vs out of scope)
  • Well-established practices (prior art)

Examples:

✅ Focused (Bootstrap-003):
"Reduce error rate through detection, diagnosis, recovery, prevention"

❌ Broad (Bootstrap-002):
"Develop test strategy" (requires scoping: what tests? which patterns? how much coverage?)

Impact: Focused scope means:

  • Less exploration needed
  • Clearer convergence criteria
  • Lower risk of scope creep

Criterion 3: Direct Validation (IMPORTANT)

Indicator: Can validate without multi-context deployment

What it means:

  • Retrospective validation possible (use historical data)
  • Single-context validation sufficient
  • Proxy metrics strongly correlate with value

Examples:

✅ Direct (Bootstrap-003):
Retrospective validation via 1,336 historical errors
No deployment needed
Confidence: 0.79

❌ Indirect (Bootstrap-002):
Multi-context validation required (3 project archetypes)
Deploy and test in each context
Adds 2-3 iterations

Impact: Direct validation means:

  • Faster iteration cycles
  • Less complexity
  • Easier V_meta calculation

See ../retrospective-validation for retrospective validation technique.

Criterion 4: Generic Agent Sufficiency (MODERATE)

Indicator: Generic agents (data-analyst, doc-writer, coder) sufficient

What it means:

  • No specialized domain knowledge required
  • Tasks are analysis + documentation + simple automation
  • Pattern extraction is straightforward

Examples:

✅ Generic sufficient (Bootstrap-003):
Generic agents analyzed errors, documented taxonomy, created scripts
No specialization overhead
3 iterations

⚠️ Specialization needed (Bootstrap-002):
coverage-analyzer (10x speedup)
test-generator (200x speedup)
6 iterations (specialization added 1-2 iterations)

Impact: No specialization means:

  • No iteration delay for agent design
  • Simpler coordination
  • Faster execution

Criterion 5: Early High-Impact Automation (MODERATE)

Indicator: Top 3 automation opportunities identified by iteration 1

What it means:

  • Pareto principle applies (20% patterns → 80% impact)
  • High-frequency, high-impact patterns obvious
  • Automation feasibility clear (no R&D risk)

Examples:

✅ Early identification (Bootstrap-003):
3 tools preventing 23.7% of errors identified in iteration 0-1
Clear automation path
Rapid V_instance improvement

⚠️ Gradual discovery (Bootstrap-002):
8 test patterns emerged gradually over 6 iterations
Pattern library built incrementally

Impact: Early automation means:

  • Faster V_instance improvement
  • Clearer path to convergence
  • Less trial-and-error

Convergence Speed Prediction Model

Formula

Predicted Iterations = Base(4) + Σ penalties

Penalties:
- V_meta(s₀) < 0.40: +2 iterations
- Domain scope fuzzy: +1 iteration
- Multi-context validation: +2 iterations
- Specialization needed: +1 iteration
- Automation unclear: +1 iteration

Worked Examples

Bootstrap-003 (Error Recovery):

Base: 4
V_meta(s₀) = 0.48 ≥ 0.40: +0 ✓
Domain scope clear: +0 ✓
Retrospective validation: +0 ✓
Generic agents sufficient: +0 ✓
Automation identified early: +0 ✓
---
Predicted: 4 iterations
Actual: 3 iterations ✅

Bootstrap-002 (Test Strategy):

Base: 4
V_meta(s₀) = 0.04 < 0.40: +2 ✗
Domain scope broad: +1 ✗
Multi-context validation: +2 ✗
Specialization needed: +1 ✗
Automation unclear: +0 ✓
---
Predicted: 10 iterations
Actual: 6 iterations ✅ (model conservative)

Interpretation: Model predicts upper bound. Actual often faster due to efficient execution.

See examples/prediction-examples.md for more cases.


Rapid Convergence Strategy

If criteria indicate 3-4 iteration potential, optimize:

Pre-Iteration 0: Planning (1-2 hours)

1. Establish Baseline Metrics

  • Identify existing data sources
  • Define quantifiable success criteria
  • Ensure automatic measurement

Example: meta-cc query-tools --status error → 1,336 errors immediately

2. Scope Domain Tightly

  • Write 1-sentence definition
  • List explicit in/out boundaries
  • Identify prior art

Example: "Error detection, diagnosis, recovery, prevention for meta-cc"

3. Plan Validation Approach

  • Prefer retrospective (historical data)
  • Minimize multi-context overhead
  • Identify proxy metrics

Example: Retrospective validation with 1,336 historical errors

Iteration 0: Comprehensive Baseline (3-5 hours)

Target: V_meta(s₀) ≥ 0.40

Tasks:

  1. Quantify current state thoroughly
  2. Create initial taxonomy (≥70% coverage)
  3. Document existing practices
  4. Identify top 3 automations

Example (Bootstrap-003):

  • Analyzed all 1,336 errors
  • Created 10-category taxonomy (79.1% coverage)
  • Documented 5 workflows, 5 patterns, 8 guidelines
  • Identified 3 tools preventing 23.7% errors
  • Result: V_meta(s₀) = 0.48 ✅

Time: Spend 3-5 hours here (saves 6-10 hours overall)

Iteration 1: High-Impact Automation (3-4 hours)

Tasks:

  1. Implement top 3 tools
  2. Expand taxonomy (≥90% coverage)
  3. Validate with data (if possible)
  4. Target: ΔV_instance = +0.20-0.30

Example (Bootstrap-003):

  • Built 3 tools (515 LOC, ~150-180 lines each)
  • Expanded taxonomy: 10 → 12 categories (92.3%)
  • Result: V_instance = 0.55 (+0.27) ✅

Iteration 2: Validate and Converge (3-4 hours)

Tasks:

  1. Test automation (real/historical data)
  2. Complete taxonomy (≥95% coverage)
  3. Check convergence:
    • V_instance ≥ 0.80?
    • V_meta ≥ 0.80?
    • System stable?

Example (Bootstrap-003):

  • Validated 23.7% error prevention
  • Taxonomy: 95.4% coverage
  • Result: V_instance = 0.83, V_meta = 0.85 ✅ CONVERGED

Total time: 10-13 hours (3 iterations)


Anti-Patterns

1. Premature Convergence

Symptom: Declare convergence at iteration 2 with V ≈ 0.75

Problem: Rushed without meeting 0.80 threshold

Solution: Rapid convergence = 3-4 iterations (not 2). Respect quality threshold.

2. Scope Creep

Symptom: Adding categories/patterns in iterations 3-4

Problem: Poorly scoped domain

Solution: Tight scoping in README. If scope grows, re-plan or accept slower convergence.

3. Over-Engineering Automation

Symptom: Spending 8+ hours on complex tools

Problem: Complexity delays convergence

Solution: Keep tools simple (1-2 hours, 150-200 lines). Complex tools are iteration 3-4 work.

4. Unnecessary Multi-Context Validation

Symptom: Testing 3+ contexts despite obvious generalizability

Problem: Validation overhead delays convergence

Solution: Use judgment. Error recovery is universal. Test strategy may need multi-context.


Comparison Table

AspectStandardRapid
Iterations5-73-4
Duration20-30h10-15h
V_meta(s₀)0.00-0.300.40-0.60
DomainBroad/exploratoryFocused
ValidationMulti-context oftenDirect/retrospective
SpecializationLikely (1-3 agents)Often unnecessary
DiscoveryIncrementalMost patterns early
RiskScope creepPremature convergence

Key: Rapid convergence is about recognizing structural factors, not rushing.


Success Criteria

Rapid convergence pattern successfully applied when:

  1. Accurate prediction: Actual iterations within ±1 of predicted
  2. Quality maintained: V_instance ≥ 0.80, V_meta ≥ 0.80
  3. Time efficiency: Duration ≤50% of standard convergence
  4. Artifact completeness: Deliverables production-ready
  5. Reusability validated: ≥80% transferability achieved

Bootstrap-003 Validation:

  • ✅ Predicted: 3-4, Actual: 3
  • ✅ Quality: V_instance=0.83, V_meta=0.85
  • ✅ Efficiency: 10h (39% of Bootstrap-002's 25.5h)
  • ✅ Artifacts: 13 categories, 8 workflows, 3 tools
  • ✅ Reusability: 85-90%

Related Skills

Parent framework:

Complementary acceleration:

Supporting:


References

Core guide:

Examples:


Status: ✅ Validated | Bootstrap-003 | 40-60% time reduction | No quality sacrifice

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Automation

subagent-prompt-construction

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

agent-observability

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

agent-evaluation

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

mcp_agent_mail

No summary provided by upstream source.

Repository SourceNeeds Review