knowledge-base-manager

Knowledge base architecture selection, curation, and governance. Use when choosing between document-based (RAG), entity-based (graph), or hybrid KB architectures, establishing content curation workflows, implementing versioning and governance, or evaluating quality metrics. For building retrieval pipelines, use the rag-implementer skill. For building knowledge graphs, use the knowledge-graph-builder skill.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "knowledge-base-manager" with this command: npx skills add oakoss/agent-skills/oakoss-agent-skills-knowledge-base-manager

Knowledge Base Manager

Overview

Provides a structured methodology for selecting, designing, and governing knowledge bases. Covers architecture decisions (document-based vs entity-based vs hybrid), content curation, quality metrics, versioning strategies, and maintenance governance. Use when choosing a KB architecture, establishing curation workflows, or building governance processes for organizational knowledge.

When NOT to use: Static documentation suffices, fewer than 50 FAQ items cover all questions, or no maintenance resources are available. For implementing retrieval pipelines (chunking, embeddings, vector stores), use the rag-implementer skill. For implementing knowledge graphs (ontology, entity extraction, graph databases), use the knowledge-graph-builder skill.

Quick Reference

AspectOptionsKey Considerations
ArchitectureDocument-based (RAG), Entity-based (Graph), HybridMatch to query patterns; start simple, add complexity when needed
Document-basedVector DB (Pinecone, Weaviate, pgvector)Best for docs, FAQs, manuals; semantic search; easy to add content
Entity-basedGraph DB (Neo4j, ArangoDB)Best for org charts, catalogs, networks; relationship traversal
HybridBoth + linking layerEnterprise, medical, legal; combined queries; highest complexity
When to skip KBStatic docs, <50 FAQ itemsNo maintenance resources, information never changes
Implementation6 phasesAudit, Curation, Storage, Quality, Versioning, Governance
Accuracy target>90% on test questionsCreate 100+ test questions with known correct answers
Coverage target>80% questions answerableValidate against real user queries continuously
Freshness target<30 days average ageAutomated freshness monitoring + scheduled updates
Consistency target>95% conflict-freeDeduplication + single source of truth
Query latency<100ms medianCaching and optimization for common access patterns
Storage techpgvector, Pinecone, Weaviate, Chromapgvector for existing Postgres; Pinecone for managed scale
Index typesHNSW, IVFFlatHNSW for recall; IVFFlat for frequently rebuilt indexes
Ingestion pipelineLoad, clean, chunk, embed, storeChunk at semantic boundaries; 512 tokens max; 10-15% overlap
DeduplicationContent hashing, semantic similarityHash for exact dupes; cosine similarity >0.95 for semantic dupes
Quality testingRecall@K, MRR, accuracy sampling100+ test questions; measure recall@10 >0.8 and MRR >0.7
Drift detectionEmbedding distribution monitoringTrack mean shift; alert when >0.1 threshold
VersioningSnapshot, Event-sourced, Git-styleSnapshot for simple; event-sourced for audit; git-style for teams
MaintenanceDaily, Weekly, Monthly, QuarterlyEstablish schedule from day 1; monitor errors and user feedback

Common Mistakes

MistakeCorrect Pattern
Ingesting raw data without curation or normalizationCurate, clean, and deduplicate before ingesting; quality over quantity
Skipping version control for KB contentImplement versioning from day one with rollback and audit trail
Building a KB without validating against user questionsStart with user research and test against real queries for >90% accuracy
Choosing hybrid architecture when document-based sufficesMatch architecture to actual query patterns; start simple, add complexity when needed
Launching without freshness monitoring or update schedulesSet up automated freshness checks and scheduled content reviews
No provenance tracking on knowledge entriesAlways track source URL, timestamp, author, and confidence score
Duplicate information across sourcesEstablish single source of truth; merge similar entries with conflict resolution rules
Perfectionism delaying launchLaunch at 80% coverage and iterate based on real usage data

Delegation

  • Audit existing knowledge sources and classify content types: Use Explore agent to inventory documents, assess quality, and identify gaps
  • Implement end-to-end KB pipeline with storage and retrieval: Use Task agent to deploy database, configure search, and run quality checks
  • Design KB architecture and governance model: Use Plan agent to select between document-based, entity-based, or hybrid approaches

For implementing document retrieval pipelines (chunking, embeddings, vector stores, hybrid search), use the rag-implementer skill. For implementing knowledge graphs (ontology design, entity extraction, graph databases), use the knowledge-graph-builder skill.

References

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Research

knowledge-graph-builder

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

playwright

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

ui-ux-polish

No summary provided by upstream source.

Repository SourceNeeds Review
Automation

tanstack-form

No summary provided by upstream source.

Repository SourceNeeds Review