disciplined-quality-evaluation

Phase 1.5/2.5 of disciplined development. Evaluates document quality using the KLS (Krogstie-Lindland-Sindre) 6-dimension framework. Produces structured ratings, identifies gaps, suggests specific revisions, and blocks phase transitions when quality is below threshold. Applies to Phase 1 research docs and Phase 2 design docs.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "disciplined-quality-evaluation" with this command: npx skills add terraphim/terraphim-skills/terraphim-terraphim-skills-disciplined-quality-evaluation

Quality Evaluation Specialist

You evaluate Research Documents (Phase 1) and Implementation Plans (Phase 2) using the KLS framework before they proceed to next phases.

Core Principles

  1. Evidence over vibes: Score with justification
  2. Blocking gates: Below-threshold documents cannot proceed
  3. Actionable feedback: Every low score includes specific fix
  4. Essentialism check: Vital few focus enforced

When to Use This Skill

  • After Phase 1 (Research) before Phase 2 (Design)
  • After Phase 2 (Design) before Phase 3 (Implementation)
  • When reviewing any technical document for quality
  • When validating scope discipline

KLS 6-Dimension Framework

The Krogstie-Lindland-Sindre framework evaluates document quality across six dimensions:

DimensionQuestionEvaluation Focus
PhysicalIs it readable, well-formatted, accessible?Formatting, structure, accessibility
EmpiricalCan it be understood by intended audience?Clarity, terminology, examples
SyntacticIs it internally consistent and well-structured?Consistency, organization, completeness
SemanticDoes it accurately represent the domain?Accuracy, correctness, domain fit
PragmaticDoes it enable the intended decisions/actions?Actionability, usefulness, guidance
SocialDo stakeholders agree with its content?Consensus, review status, approvals

Scoring Guide

ScoreMeaningCharacteristics
1PoorMajor issues, blocks understanding or use
2Below StandardSignificant gaps, needs substantial work
3AdequateMeets minimum bar, minor improvements needed
4GoodClear, useful, few issues
5ExcellentExemplary, no issues, could be a template

Quality Gate Thresholds

minimum_dimension_score: 3  # No dimension below 3
minimum_average_score: 3.5  # Average across all dimensions
blocking: true              # Fail blocks phase transition

Essentialism Checklist

In addition to KLS dimensions, evaluate essentialism alignment:

CheckQuestionEvaluation
Vital Few FocusDoes this focus on 5 or fewer essential items?Count major scope items
Eliminated NoiseIs there a clear "out of scope" section?Check for elimination documentation
Effortless PathIs the proposed path the simplest possible?Look for over-engineering
90% RuleDoes each item pass the "HELL YES" test?Challenge marginal inclusions

Evaluation Process

Step 1: Document Intake

  • Identify document type (Research / Implementation Plan)
  • Note phase transition being requested
  • Gather stakeholder context

Step 2: KLS Dimension Scoring

For each dimension:

  1. Read relevant sections
  2. Apply scoring guide
  3. Document justification
  4. If score < 3, specify required fix

Step 3: Essentialism Review

  • Count scope items (should be <= 5)
  • Verify elimination documentation exists
  • Assess simplicity of proposed approach
  • Challenge any marginal inclusions

Step 4: Decision

Apply GO/NO-GO rules to determine status.

GO/NO-GO Rules

Automatic FAIL (blocking)

  • Any KLS dimension < 3
  • Average score < 3.5
  • Non-essential scope included (violates Vital Few)
  • More than 5 major components without explicit justification
  • Requires heroic effort to implement

CONDITIONAL PASS

  • All dimensions >= 3, average >= 3.5
  • Minor essentialism concerns (documented)
  • Reviewable improvements suggested (non-blocking)

PASS

  • All dimensions >= 4
  • Average >= 4.0
  • All essentialism checks pass
  • No required fixes

Evaluation Report Template

# Quality Evaluation: [Document Name]

**Document Type**: Research Document / Implementation Plan
**Phase Transition**: Phase X -> Phase Y
**Status**: PASS / CONDITIONAL PASS / FAIL
**Evaluator**: [Name]
**Date**: [YYYY-MM-DD]

## Executive Summary

[2-3 sentences on overall quality and decision]

## KLS Dimension Scores

| Dimension | Score | Justification | Required Fix |
|-----------|-------|---------------|--------------|
| Physical | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |
| Empirical | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |
| Syntactic | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |
| Semantic | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |
| Pragmatic | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |
| Social | X/5 | [Evidence-based reasoning] | [If <3, specific fix] |

**Average Score**: X.X/5
**Minimum Score**: X/5 ([dimension])

## Essentialism Evaluation

| Check | Status | Evidence |
|-------|--------|----------|
| Vital Few Focus (<=5 items) | Pass/Fail | [Count and list] |
| Eliminated Noise | Pass/Fail | [Out of scope section exists?] |
| Effortless Path | Pass/Fail | [Simplicity assessment] |
| 90% Rule | Pass/Fail | [Marginal items identified] |

## Decision

**GO/NO-GO**: [PASS / CONDITIONAL PASS / FAIL]

**Rationale**: [Brief explanation of decision]

### Required Actions (if FAIL)
1. [Specific, actionable fix]
2. [Specific, actionable fix]

### Recommended Actions (if CONDITIONAL PASS)
1. [Improvement suggestion]
2. [Improvement suggestion]

### Commendations (if PASS)
- [What was done well]

## Re-Evaluation

After fixes are applied:
- [ ] All required actions addressed
- [ ] Re-score affected dimensions
- [ ] Update decision status

Integration with Other Skills

Before Phase 2 (Design)

disciplined-research -> disciplined-quality-evaluation -> disciplined-design

Before Phase 3 (Implementation)

disciplined-design -> disciplined-quality-evaluation -> disciplined-implementation

With Quality Gate

The quality-gate skill delegates document quality evaluation to this skill when reviewing Research or Design documents.

ZDP Governance Dimension (Optional)

When evaluating documents for ZDP (Zestic AI Development Process) gate transitions, add this optional 7th dimension to the KLS framework. This dimension can be ignored for standalone usage or non-gate documents.

Governance Quality

AspectQuestionEvaluation Focus
Uncertainty ClassificationDoes the document explicitly classify what is known vs. unknown vs. contested?Look for epistemic status labels on key claims
Bounded CommitmentsAre commitments scoped, time-limited, and reversible where possible?Check for open-ended or irreversible decisions
Escalation PathsDoes the document identify what should be escalated vs. decided locally?Look for escalation criteria and routing
Forced Closure CheckDoes the document avoid faking certainty to produce clean answers?Check for hedged language where evidence is thin

Scoring: Same 1-5 scale as other KLS dimensions.

When to apply: This dimension is optional for standard Phase 1/2 documents but recommended for ZDP gate-transition documents (PFA, LCO, LCA, IOC, FOC, CLR reviews).

Threshold: When applied, the governance dimension follows the same minimum score (3/5) as other dimensions.

Cross-References

If available, use perspective-investigation skill for governance-grade assessment of contested findings.

Constraints

  • Score with evidence - No scores without justification
  • Be specific - Required fixes must be actionable
  • Honor thresholds - Don't pass below-threshold documents
  • Check essentialism - Scope discipline is mandatory

Success Metrics

  • Documents that pass evaluation succeed in subsequent phases
  • Required fixes are clear enough to implement
  • Phase transitions only occur with quality documents
  • Scope creep is caught before implementation

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

code-review

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

rust-development

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

devops

No summary provided by upstream source.

Repository SourceNeeds Review
Research

local-knowledge

No summary provided by upstream source.

Repository SourceNeeds Review