Options Comparator
Structured frameworks for systematically comparing alternatives, scoring options, and producing defensible recommendations.
Weighted Scoring Matrix
Standard Weighted Matrix Template
WEIGHTED SCORING MATRIX:
STEP 1: Define criteria and weights (must sum to 100%)
| Criterion | Weight | Option A | Option B | Option C |
|-----------------|--------|----------|----------|----------|
| [Criterion 1] | 25% | [1-5] | [1-5] | [1-5] |
| [Criterion 2] | 20% | [1-5] | [1-5] | [1-5] |
| [Criterion 3] | 20% | [1-5] | [1-5] | [1-5] |
| [Criterion 4] | 15% | [1-5] | [1-5] | [1-5] |
| [Criterion 5] | 10% | [1-5] | [1-5] | [1-5] |
| [Criterion 6] | 10% | [1-5] | [1-5] | [1-5] |
|-----------------|--------|----------|----------|----------|
| WEIGHTED TOTAL | 100% | [sum] | [sum] | [sum] |
STEP 2: Calculate weighted scores
Weighted score = Raw score x Weight
Total = Sum of all weighted scores
STEP 3: Interpret results
4.5-5.0: Excellent fit
3.5-4.4: Good fit
2.5-3.4: Acceptable with trade-offs
1.5-2.4: Poor fit — significant concerns
1.0-1.4: Disqualified
SCORING RUBRIC:
5 = Exceeds requirements / best in class
4 = Fully meets requirements
3 = Partially meets requirements
2 = Significant gaps
1 = Does not meet requirements / disqualifying
Weight Assignment Methods
| Method | How It Works | Best For |
|---|
| Direct assignment | Stakeholders allocate 100 points across criteria | Small groups, quick decisions |
| Pairwise comparison | Compare criteria two at a time, derive weights | Rigorous prioritization |
| MoSCoW ranking | Must/Should/Could/Won't, then assign within tiers | Requirements-driven decisions |
| Swing weighting | Rate criteria by how much best-to-worst matters | Complex multi-attribute decisions |
| Stakeholder voting | Each stakeholder distributes 10 votes | Democratic team decisions |
Weight Validation Checklist
BEFORE FINALIZING WEIGHTS:
- [ ] Weights sum to exactly 100%
- [ ] No single criterion exceeds 40% (unless justified)
- [ ] No criterion is below 5% (drop it if irrelevant)
- [ ] Weights reflect stated priorities (not just habit)
- [ ] Stakeholders reviewed and approved weights
- [ ] Weights were set BEFORE scoring options
(prevents reverse-engineering to a preferred choice)
Pairwise Comparison
Pairwise Comparison Matrix
PAIRWISE COMPARISON TEMPLATE:
Compare criteria A through E. For each pair, indicate
which is more important (mark the winner):
A B C D E WINS WEIGHT
A [--] [ ] [A] [ ] [A] 2 25%
B [B] [--] [B] [B] [B] 4 40%
C [ ] [ ] [--] [ ] [C] 1 10%
D [D] [ ] [D] [--] [D] 3 25%
E [ ] [ ] [ ] [ ] [--] 0 0%
Weight = Wins / Total comparisons x 100
Total comparisons = n(n-1)/2 = 5(4)/2 = 10
INSTRUCTIONS:
1. Compare each pair: "Is criterion X more important than Y?"
2. Mark the winner in the matrix
3. Count wins for each criterion
4. Calculate weights from win percentages
5. Adjust if any criterion has 0% but should remain
Forced Ranking
FORCED RANKING METHOD:
List all options and rank from best to worst on each criterion.
No ties allowed (forces differentiation).
| Criterion | Rank 1 (Best) | Rank 2 | Rank 3 | Rank 4 (Worst) |
|---------------|---------------|--------|--------|-----------------|
| Price | Option C | Option A| Option D| Option B |
| Quality | Option B | Option D| Option A| Option C |
| Speed | Option A | Option B| Option C| Option D |
| Support | Option D | Option C| Option B| Option A |
SCORING:
Rank 1 = 4 points, Rank 2 = 3, Rank 3 = 2, Rank 4 = 1
(Or weight the ranking scores by criterion importance)
Pros/Cons with Weights
Structured Pros/Cons Template
WEIGHTED PROS/CONS ANALYSIS:
OPTION: [Name]
PROS:
| # | Advantage | Impact | Certainty | Score |
|---|------------------------------|--------|-----------|-------|
| 1 | [Pro description] | H/M/L | H/M/L | [1-9] |
| 2 | [Pro description] | H/M/L | H/M/L | [1-9] |
| 3 | [Pro description] | H/M/L | H/M/L | [1-9] |
CONS:
| # | Disadvantage | Impact | Certainty | Score |
|---|------------------------------|--------|-----------|-------|
| 1 | [Con description] | H/M/L | H/M/L | [1-9] |
| 2 | [Con description] | H/M/L | H/M/L | [1-9] |
| 3 | [Con description] | H/M/L | H/M/L | [1-9] |
SCORING GUIDE:
Impact: High=3, Medium=2, Low=1
Certainty: High=3, Medium=2, Low=1
Score = Impact x Certainty (range: 1-9)
NET SCORE = Sum of Pro scores - Sum of Con scores
Positive: Pros outweigh cons
Negative: Cons outweigh pros
Near zero: Trade-off decision (needs judgment)
Comparative Pros/Cons
| Factor | Option A | Option B | Option C |
|---|
| Best for | [ideal use case] | [ideal use case] | [ideal use case] |
| Worst for | [poor fit scenario] | [poor fit scenario] | [poor fit scenario] |
| Top Pro | [strongest advantage] | [strongest advantage] | [strongest advantage] |
| Top Con | [biggest drawback] | [biggest drawback] | [biggest drawback] |
| Risk level | Low / Medium / High | Low / Medium / High | Low / Medium / High |
| Reversibility | Easy / Hard / Impossible | Easy / Hard / Impossible | Easy / Hard / Impossible |
Decision Matrix Template
Comprehensive Decision Matrix
DECISION MATRIX:
DECISION: [Clear statement of what you are deciding]
DATE: [Date of analysis]
OWNER: [Decision maker(s)]
DEADLINE: [When decision must be made]
OPTIONS UNDER CONSIDERATION:
A. [Option name and brief description]
B. [Option name and brief description]
C. [Option name and brief description]
D. [Status quo / do nothing]
MUST-HAVE CRITERIA (pass/fail — eliminates options):
| Requirement | Option A | Option B | Option C | Option D |
|----------------------|----------|----------|----------|----------|
| [Hard requirement 1] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|
| [Hard requirement 2] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|
| [Hard requirement 3] | Pass/Fail| Pass/Fail| Pass/Fail| Pass/Fail|
NICE-TO-HAVE CRITERIA (scored and weighted):
| Criterion | Weight | Opt A | Opt B | Opt C | Opt D |
|-------------|--------|-------|-------|-------|-------|
| [Criterion] | X% | [1-5] | [1-5] | [1-5] | [1-5] |
| [Criterion] | X% | [1-5] | [1-5] | [1-5] | [1-5] |
| [Criterion] | X% | [1-5] | [1-5] | [1-5] | [1-5] |
|-------------|--------|-------|-------|-------|-------|
| TOTAL | 100% | [sum] | [sum] | [sum] | [sum] |
RECOMMENDATION: [Option letter] because [1-2 sentence rationale]
RISKS OF CHOSEN OPTION:
1. [Risk and mitigation plan]
2. [Risk and mitigation plan]
NEXT STEPS:
1. [Action item, owner, deadline]
2. [Action item, owner, deadline]
Trade-Off Analysis Framework
Trade-Off Mapping
TRADE-OFF ANALYSIS:
STEP 1: Identify the key trade-off dimensions
Common trade-offs:
- Cost vs Quality
- Speed vs Thoroughness
- Flexibility vs Standardization
- Control vs Convenience
- Short-term vs Long-term
- Risk vs Reward
- Simplicity vs Capability
STEP 2: Map options on trade-off axes
HIGH QUALITY
|
| Option B
| *
|
LOW COST -------+--------- HIGH COST
|
Option A|
* |
| Option C
| *
LOW QUALITY
STEP 3: Identify the efficient frontier
Options on the frontier are rationally competitive.
Options below the frontier are dominated
(another option is better on all axes).
STEP 4: Choose based on priorities
"We are optimizing for [dimension] while keeping
[other dimension] above [minimum threshold]."
Trade-Off Decision Rules
| Rule | When to Use | How It Works |
|---|
| Maximize one, threshold others | Clear primary objective | Set minimums for secondary criteria, then pick highest on primary |
| Satisfice | Time-pressured, good enough is fine | Pick first option that meets all minimum thresholds |
| Lexicographic | Clear priority ordering | Sort by most important criterion first, break ties with second |
| Minimax regret | High uncertainty | Choose option that minimizes worst-case disappointment |
| Expected value | Quantifiable outcomes and probabilities | Probability x payoff for each scenario, pick highest EV |
Sensitivity Analysis for Decisions
Weight Sensitivity Testing
SENSITIVITY ANALYSIS:
PURPOSE: Test if the recommendation changes when weights shift.
BASELINE WEIGHTS:
Cost: 30% | Quality: 25% | Speed: 20% | Support: 15% | Risk: 10%
Winner: Option B (score: 3.85)
SCENARIO 1 — Cost-focused (Cost +15%, others proportionally reduced):
Cost: 45% | Quality: 20% | Speed: 16% | Support: 12% | Risk: 7%
Winner: [recalculate]
SCENARIO 2 — Quality-focused (Quality +15%):
Cost: 24% | Quality: 40% | Speed: 16% | Support: 12% | Risk: 8%
Winner: [recalculate]
SCENARIO 3 — Risk-averse (Risk +20%):
Cost: 22% | Quality: 19% | Speed: 15% | Support: 14% | Risk: 30%
Winner: [recalculate]
INTERPRETATION:
If the same option wins in all scenarios → ROBUST decision
If winner changes in 1 scenario → Note the sensitivity
If winner changes in 2+ scenarios → Decision depends on priorities
Score Sensitivity Testing
BREAKEVEN ANALYSIS:
"How much would Option A's score on [criterion] need
to improve to overtake Option B?"
Current:
Option A total: 3.45
Option B total: 3.85
Gap: 0.40
Criterion X (weight 25%):
Option A score: 2
Required score to close gap: 2 + (0.40 / 0.25) = 3.6 → round to 4
Is this plausible? [Yes/No]
If yes → decision is sensitive to this criterion
If no → decision is robust on this dimension
Recommendation Memo Template
Executive Decision Memo
DECISION RECOMMENDATION MEMO
TO: [Decision maker(s)]
FROM: [Analyst / Team]
DATE: [Date]
RE: Recommendation: [Decision topic]
─────────────────────────────────────────────
EXECUTIVE SUMMARY:
We recommend [Option X] for [one-sentence rationale].
This option scores highest across our evaluation criteria,
particularly in [top 2 criteria]. Estimated [cost/timeline]:
[key number]. Key risk: [top risk and mitigation].
─────────────────────────────────────────────
BACKGROUND:
[2-3 sentences on why this decision is needed now]
OPTIONS EVALUATED:
A. [Option and one-line description]
B. [Option and one-line description]
C. [Option and one-line description]
EVALUATION CRITERIA AND WEIGHTS:
[Criterion 1] (X%) | [Criterion 2] (X%) | [Criterion 3] (X%)
SCORING SUMMARY:
| Option | Score | Rank | Key Strength | Key Weakness |
|--------|-------|------|----------------------|---------------------|
| A | 3.45 | 2 | [strength] | [weakness] |
| B | 3.85 | 1 | [strength] | [weakness] |
| C | 2.90 | 3 | [strength] | [weakness] |
RECOMMENDATION: Option B
Rationale: [3-5 sentences explaining why, addressing trade-offs]
SENSITIVITY: This recommendation holds under all tested scenarios
except [edge case], which would require [condition].
RISKS AND MITIGATIONS:
1. [Risk]: [Mitigation plan]
2. [Risk]: [Mitigation plan]
IMPLEMENTATION PLAN:
1. [Step, owner, date]
2. [Step, owner, date]
3. [Decision review checkpoint, date]
─────────────────────────────────────────────
APPENDIX: Detailed scoring matrix, sensitivity analysis
Vendor Evaluation Scorecard
Vendor Assessment Template
VENDOR EVALUATION SCORECARD:
VENDOR: [Company name]
EVALUATED BY: [Names]
DATE: [Date]
PRODUCT/SERVICE: [What you are evaluating]
CATEGORY 1: PRODUCT FIT (30% weight)
| Criterion | Score (1-5) | Notes |
|----------------------------|-------------|----------------------|
| Feature completeness | | |
| Integration capability | | |
| Scalability | | |
| Customization options | | |
| User experience / UI | | |
| Category subtotal | /25 | |
CATEGORY 2: COMMERCIAL (25% weight)
| Criterion | Score (1-5) | Notes |
|----------------------------|-------------|----------------------|
| Total cost of ownership | | |
| Pricing transparency | | |
| Contract flexibility | | |
| Payment terms | | |
| ROI timeline | | |
| Category subtotal | /25 | |
CATEGORY 3: SUPPORT AND SERVICE (20% weight)
| Criterion | Score (1-5) | Notes |
|----------------------------|-------------|----------------------|
| Implementation support | | |
| Training resources | | |
| Ongoing customer support | | |
| SLA commitments | | |
| Account management | | |
| Category subtotal | /25 | |
CATEGORY 4: COMPANY VIABILITY (15% weight)
| Criterion | Score (1-5) | Notes |
|----------------------------|-------------|----------------------|
| Financial stability | | |
| Market position | | |
| Product roadmap | | |
| Customer references | | |
| Industry reputation | | |
| Category subtotal | /25 | |
CATEGORY 5: RISK (10% weight)
| Criterion | Score (1-5) | Notes |
|----------------------------|-------------|----------------------|
| Data security / compliance | | |
| Vendor lock-in risk | | |
| Migration complexity | | |
| Business continuity plan | | |
| Reference check results | | |
| Category subtotal | /25 | |
OVERALL WEIGHTED SCORE: [calculated] / 5.0
RECOMMENDATION: Proceed / Shortlist / Reject
Technology Selection Framework
Technology Evaluation Criteria
TECHNOLOGY SELECTION MATRIX:
FUNCTIONAL FIT:
- [ ] Meets core requirements (pass/fail list)
- [ ] Handles expected scale (users, data, transactions)
- [ ] Integrates with existing stack
- [ ] Supports required platforms/environments
DEVELOPER EXPERIENCE:
- [ ] Documentation quality and completeness
- [ ] Community size and activity (GitHub stars, forums)
- [ ] Learning curve for the team
- [ ] Tooling and IDE support
- [ ] Error messages and debugging experience
OPERATIONAL:
- [ ] Deployment model fits infrastructure
- [ ] Monitoring and observability support
- [ ] Backup and disaster recovery
- [ ] Security track record and patching cadence
STRATEGIC:
- [ ] Aligned with technology direction
- [ ] Vendor/project longevity (not abandonware)
- [ ] Hiring market (can you find people who know it?)
- [ ] Exit strategy (migration path if you switch later)
TOTAL COST:
- [ ] License / subscription fees
- [ ] Infrastructure costs
- [ ] Training and ramp-up time
- [ ] Maintenance and operations
- [ ] Opportunity cost of alternatives
Build vs Buy Decision
| Factor | Build | Buy | Hybrid |
|---|
| Core differentiator? | Yes — build it | No — buy it | Customize a platform |
| Team has expertise? | Yes | No | Partial |
| Time to value | Months | Weeks | Weeks-Months |
| Long-term cost | Higher (maintenance) | Predictable (subscription) | Mixed |
| Control | Full | Limited | Moderate |
| Risk | Technical debt | Vendor dependency | Both |
| Best when | Unique requirements, strategic IP | Commodity functionality | 80/20 fit |
Decision Anti-Patterns
COMMON DECISION MISTAKES:
1. ANALYSIS PARALYSIS
Symptom: Endless evaluation, no decision made
Fix: Set a decision deadline and "good enough" threshold
2. ANCHORING TO FIRST OPTION
Symptom: First option evaluated becomes the default
Fix: Evaluate all options before scoring any
3. CONFIRMATION BIAS
Symptom: Seeking data that supports preferred option
Fix: Assign a devil's advocate for each option
4. SUNK COST FALLACY
Symptom: Sticking with an option because of past investment
Fix: Evaluate options on future value only
5. RECENCY BIAS
Symptom: Overweighting the last demo or reference call
Fix: Standardize evaluation timing and criteria
6. GROUPTHINK
Symptom: Team converges without genuine debate
Fix: Independent scoring before group discussion
7. FEATURE COUNTING
Symptom: Most features = best option (ignoring fit)
Fix: Weight criteria by importance, not count
8. IGNORING STATUS QUO
Symptom: Not comparing options against doing nothing
Fix: Always include "do nothing" as Option D
See Also