trusted-skill-scout

Discover, trust-filter, and install useful skills for any repository. Use this whenever a user asks to find skills, compare skill options, or install skills safely for their current project, even if they do not mention "trust" or "security" explicitly.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "trusted-skill-scout" with this command: npx skills add <owner/repo>

Trusted Skill Scout

Find high-signal skills for the current repository, filter by trust policy, collect user selections, and install deterministically.

Core Behavior

Follow this flow in order. Do not skip gates.

  1. Ask install scope before discovery (default: project-only), then state what this skill does.
  2. Profile the repo (parallel sub-agents when available).
  3. Generate 6-10 repo-aware discovery queries.
  4. Run discovery queries sequentially with npx skills find.
  5. Enrich candidates with trust metadata.
  6. Apply trust filter.
  7. Rank, present compact cards, and collect keep selections (interactive multi-select when available).
  8. Build final shortlist and ask install confirmation.
  9. Install selected skills sequentially with deterministic commands.
  10. Verify and report results.

Step 0: Ask Scope First

Call the Question tool exactly once at the beginning:

Question: "Install approved skills as Global or Project-only? (default: Project-only)"

After the user responds, state the purpose clearly in one line before proceeding:

What this skill does: discovers repo-relevant skills, applies a trust gate, and installs only the skills you explicitly approve.

Rules:

  • If the user does not answer explicitly or says "default", use Project-only.
  • Persist this choice for all install commands in this run.
  • Do not run any discovery steps before the user has answered this question.

Step 1: Profile Repository

Preferred mode: parallel sub-agents

If the Task tool is available, spawn sub-agents in parallel:

  • Sub-agent A: stack and runtime detection (languages, frameworks, package managers).
  • Sub-agent B: workflow detection (test, lint, build, deploy, CI, release).
  • Sub-agent C: product intent and priorities (README, docs, issues, architecture clues).

Collect and merge into a single repo profile:

  • stack: key technologies
  • workflows: key engineering workflows
  • goals: likely user goals and pain points
  • constraints: notable constraints (legacy, monorepo, compliance, etc.)

Fallback mode: serial profiling

If sub-agents are unavailable, run the same profiling dimensions serially.

Step 2: Generate Discovery Queries (6-10)

Generate 6-10 specific queries based on the repo profile.

Good query dimensions:

  • framework best practices
  • testing and QA
  • deployment and CI/CD
  • architecture/refactoring
  • performance and observability
  • domain-specific workflows (auth, payments, data pipelines, mobile release, etc.)

Do not use a fixed redesign-only query list.

Step 3: Run Discovery One Query at a Time

Critical: Do NOT batch or parallelize the bash calls. Run one query, present results to the user, collect their selection via the Question tool, then move to the next query.

For each query, the sequence is:

  1. Run: npx skills find "<query>"
  2. Apply trust filter (Step 5) and fit scoring (Step 6) to the results.
  3. Present the ranked cards for that query (see Step 7).
  4. Call the Question tool to collect the user's selection before running the next query.
  5. Repeat for the next query.

From each command output, extract candidate rows with:

  • owner/repo@skill_name
  • installs (if present)
  • skills.sh URL

Normalization rules:

  • Strip ANSI color escape sequences before parsing.
  • Parse installs as an integer; K/M suffixes should be converted to absolute values (e.g. 8.1K8100).
  • If installs missing, set installs to 0.

Step 4: Enrich Candidate Metadata

For each candidate, collect:

  • owner
  • repo
  • skill_name (the exact token after @ from discovery output)
  • skills_url
  • installs
  • stars
  • owner_type if available

Recommended enrichment order:

  1. Fetch skills_url and parse repository link and stars.
  2. If stars still missing, query GitHub repo metadata.

If stars missing after enrichment, set stars to 0.

Step 5: Trust Filter (Hard Gate)

Use policy from references/trust-policy.md.

A candidate is eligible only if at least one is true:

  • owner is allowlisted
  • installs >= 100
  • stars >= 500

Missing installs/stars count as 0 unless allowlisted. Candidates that fail are excluded from user options.

Step 6: Fit Scoring and Ranking

Use rubric from references/scoring-rubric.md.

Score each eligible candidate across:

  • Stack Fit (1-5)
  • Expected Impact (1-5)
  • Overlap Risk (1-5; lower is better)
  • Implementation Clarity (1-5)

Compute:

  • overlap_bonus = 6 - overlap_risk
  • fit_total = stack_fit + expected_impact + overlap_bonus + implementation_clarity (max 20)

Verdict:

  • Strong fit if fit_total >= 15
  • Maybe if fit_total >= 11
  • otherwise exclude

Sort with tie-breakers (if fit is equal, prioritize popularity):

  1. fit_total desc
  2. installs desc
  3. stars desc
  4. owner/repo asc
  5. skill_name asc

Step 7: Present Per-Query Options

For each query, show only trust-eligible candidates as compact cards.

Display count rules:

  • Show up to 4 ranked candidates per query.
  • Target 3-4 options when available.
  • If fewer than 3 are eligible, show all eligible options.

Card format (use this exact layout every time, no variation):

[N] owner/repo@skill_name
Installs: <count> | Stars: <count>
Fit: <fit_total>/20 | Verdict: <Strong fit | Maybe>
Why: <one line repo-specific rationale>
Agent recommendation: <Keep | Skip> — <short reason>
  • Always show the raw parsed counts on the Installs/Stars line, even if one is 0.
  • Do not add a separate Trust line — the numbers speak for themselves.

After presenting all cards for the query, call the Question tool:

Question: "Query N/total — Select skills to keep (e.g. 1,3 or none):"

After receiving the answer, echo a deterministic record line before moving on:

Selection recorded: Keep: 1,3

or if nothing selected:

Selection recorded: Keep: none

Selection rules:

  • Accept only indices shown for that query.
  • Do not proceed to the next query until the user has answered.

Step 8: Final Shortlist

After all queries:

  • dedupe selections by (owner, repo, skill_name)
  • provide overlap/conflict notes
  • propose a best bundle (usually 3-5 complementary skills)
  • call the Question tool for the final gate:
Question: "Install these now? (Yes / Revise list)"

Proceed only if the user answers Yes or equivalent confirmation. If they say Revise, ask which skills to add or remove and update the shortlist before asking again.

Step 9: Install Sequentially (Deterministic)

Install each selected skill sequentially.

Project-only scope:

npx skills add <owner/repo> --skill "<skill_name>" -y

Global scope:

npx skills add <owner/repo> --skill "<skill_name>" -g -y

If one install fails:

  • log failure
  • continue remaining installs
  • provide exact retry command for failed item

Step 10: Verify and Report

Verification commands:

  • Project-only: npx skills list
  • Global: npx skills ls -g
  • Update health: npx skills check

Return final report:

  • installed successfully
  • failed installs with likely cause
  • retry commands
  • optional maintenance command: npx skills update

Output Template

Use concise sections:

  1. Scope
  • selected scope
  • whether default was used
  • one-line purpose statement shown to user
  1. Repo Profile
  • stack
  • workflows
  • goals
  1. Query N Results
  • ranked eligible cards (up to 4 per query)
  • one explicit agent recommendation per card
  • Question tool call for user selection before next query
  1. Final Shortlist
  • deduped list
  • overlap notes
  • recommended bundle
  • install confirmation prompt
  1. Installation Report
  • per-skill result
  • verification summary
  • next maintenance command

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

compliance-evidence-assembler

把审计所需证据整理成目录、清单和缺失项,便于后续评审。;use for compliance, evidence, audit workflows;do not use for 伪造证据, 替代正式审计结论.

Archived SourceRecently Updated
Security

skillguard-hardened

Security guard for OpenClaw skills, developed and maintained by rose北港(小红帽 / 猫猫帽帽). Audits installed or incoming skills with local rules plus Zenmux AI intent review, then recommends pass, warn, block, or quarantine.

Archived SourceRecently Updated
Security

api-contract-auditor

审查 API 文档、示例和字段定义是否一致,输出 breaking change 风险。;use for api, contract, audit workflows;do not use for 直接改线上接口, 替代契约测试平台.

Archived SourceRecently Updated
Security

ai-workflow-red-team-lite

对 AI 自动化流程做轻量红队演练,聚焦误用路径、边界失败和数据泄露风险。;use for red-team, ai, workflow workflows;do not use for 输出可直接滥用的攻击脚本, 帮助破坏系统.

Archived SourceRecently Updated