Agent Memory Systems (PostgreSQL)
Persistent shared memory for all AI agents. PostgreSQL 14+ on Linux or Windows. Memory failures look like intelligence failures — this skill ensures the right memory is retrieved at the right time.
Quick Start
Database agent_memory and all functions are created by init.sql in this skill directory.
Linux
psql -U postgres -c "CREATE DATABASE agent_memory;" psql -U postgres -d agent_memory -f init.sql
Windows (adjust path to your psql.exe)
& "C:\Program Files\PostgreSQL\18\bin\psql.exe" -U postgres -c "CREATE DATABASE agent_memory;" & "C:\Program Files\PostgreSQL\18\bin\psql.exe" -U postgres -d agent_memory -f init.sql
Verify: SELECT * FROM memory_health_check();
Pure Skill Mode (default)
This skill works without installing any plugin. In pure skill mode:
-
you manually run scripts when you want (progressive disclosure)
-
no global OpenCode config is modified automatically
Optional bootstrap (asks + records choices + tries to install)
Notes:
-
Interactive mode defaults to NOT installing heavy optional components.
-
Use -InstallAll / --install-all only when you're ready to install everything.
Run the bootstrap script to choose optional components (pgpass, local embeddings, pgvector) and record decisions.
Windows:
run from the skill directory
powershell.exe -NoProfile -ExecutionPolicy Bypass -File "scripts\bootstrap.ps1"
Linux/macOS:
run from the skill directory
bash "scripts/bootstrap.sh"
The selection record is stored at:
- ~/.config/opencode/agent-memory-systems-postgres/setup.json
Agent rule:
-
If this file does not exist, ask the user if they want to enable optional components.
-
Recommended: run bootstrap with all options enabled (then fix any failures it reports).
On Windows, pgvector installation follows the official pgvector instructions (Visual Studio C++ + nmake /F Makefile.win ). The bootstrap will attempt to install prerequisites via winget .
Optional automation: compaction logging (OpenCode plugin)
If you want automatic compaction logging, install the OpenCode plugin template shipped with this skill.
-
Copy plugins/agent-memory-systems-postgres.js to ~/.config/opencode/plugins/
-
Restart OpenCode
Credentials (psql)
Do NOT hardcode passwords in scripts, skill docs, or config files.
Recommended options for non-interactive psql :
-
.pgpass / pgpass.conf (recommended)
-
Linux/macOS: ~/.pgpass (must be chmod 0600 ~/.pgpass or libpq will ignore it)
-
Windows: %APPDATA%\postgresql\pgpass.conf (example: C:\Users<you>\AppData\Roaming\postgresql\pgpass.conf )
-
Format: hostname:port:database:username:password
-
Docs: https://www.postgresql.org/docs/current/libpq-pgpass.html
-
PGPASSFILE (optional override): point to a custom location for the password file
-
PGPASSWORD (not recommended): only for quick local testing; environment variables can leak on some systems
-
Docs: https://www.postgresql.org/docs/current/libpq-envars.html
Tip: set connection defaults once (per shell) to shorten commands:
export PGHOST=localhost export PGPORT=5432 export PGDATABASE=agent_memory export PGUSER=postgres
One-time setup helper scripts
This skill ships helper scripts (relative paths):
-
scripts/setup-pgpass.ps1
-
scripts/setup-pgpass.sh
OpenCode usage: run them from the skill directory.
Windows run:
powershell.exe -NoProfile -ExecutionPolicy Bypass -File "scripts\setup-pgpass.ps1"
Linux/macOS run:
bash "scripts/setup-pgpass.sh"
Memory Types
Type Lifespan Use When
working
24h auto-expire Current conversation context (requires session_id )
episodic
Permanent + decay Problem-solving experiences, debugging sessions
semantic
Permanent Extracted facts, knowledge, patterns
procedural
Permanent Step-by-step procedures, checklists (importance >= 7)
Core Functions
store_memory(type, category, tags[], title, content, metadata, agent_id, session_id, importance)
Auto-deduplicates by content hash. Duplicate inserts bump access_count and importance_score .
SELECT store_memory( 'semantic', 'windows-networking', ARRAY['ssh', 'tunnel', 'port-conflict'], 'SSH Tunnel Port Conflict Resolution', 'Fix: 1) taskkill /F /IM ssh.exe 2) Use processId not pid 3) Wait 3s', '{"os": "Windows 11"}', 'sisyphus', NULL, 9.0 );
search_memories(query, types[], categories[], tags[], agent_id, min_importance, limit)
Hybrid search: full-text (tsvector) + trigram similarity (pg_trgm) + tag filtering. Accepts plain English queries — no tsquery syntax needed. Relevance scoring: text_score * decay * recency * importance .
-- Natural language SELECT * FROM search_memories('ssh tunnel port conflict', NULL, NULL, NULL, NULL, 7.0, 5);
-- Filter by type + tags SELECT * FROM search_memories( 'troubleshooting steps', ARRAY['procedural']::memory_type[], NULL, ARRAY['ssh'], NULL, 0.0, 5 );
Returns: id, memory_type, category, title, content, importance_score, relevance_score, match_type
Where match_type is one of: fulltext , trigram_title , trigram_content , metadata .
memory_health_check()
Returns: metric | value | status for total_memories , avg_importance , stale_count .
apply_memory_decay()
Decays episodic memories by 0.9999^days_since_access . Run daily.
prune_stale_memories(age_days, max_importance, max_access_count)
Soft-deletes old episodic memories below thresholds. Default: 180 days, importance <= 3, never accessed.
Agent Workflow
Before a task
SELECT id, title, content, relevance_score FROM search_memories('keywords from user request', NULL, NULL, NULL, NULL, 5.0, 5);
If relevant memories found, reference them: "Based on past experience (memory #1)..."
After solving a problem
SELECT store_memory( 'semantic', 'category-name', ARRAY['tag1', 'tag2', 'tag3'], 'One-line problem summary', 'Detailed problem + solution', '{"os": "...", "tools": [...]}', 'agent-name', NULL, 8.0 );
When delegating to subagents
Include in prompt:
MUST DO FIRST: Search agent_memories: SELECT * FROM search_memories('relevant keywords', NULL, NULL, NULL, NULL, 5.0, 5);
MUST DO AFTER: If you solved something new, store it with store_memory(...)
Task Memory Layer (optional)
This skill also ships a minimal task/issue layer inspired by Beads: graph semantics + deterministic "ready work" queries.
Objects:
-
agent_tasks : tasks (status, priority, assignee)
-
task_links : typed links (blocks , parent_child , related , etc.)
-
blocked_tasks_cache : materialized cache to make ready queries fast
-
task_memory_links : link tasks to memories (agent_memories ) for outcomes/notes
Create tasks:
INSERT INTO agent_tasks(title, description, created_by, priority) VALUES ('Install pgvector', 'Windows build + enable extension', 'user', 1);
Add dependencies:
-- Task 1 blocks task 2 INSERT INTO task_links(from_task_id, to_task_id, link_type) VALUES (1, 2, 'blocks');
-- Task 2 is parent of task 3 (used for transitive blocking) INSERT INTO task_links(from_task_id, to_task_id, link_type) VALUES (2, 3, 'parent_child');
Rebuild blocked cache (usually auto via triggers):
SELECT rebuild_blocked_tasks_cache();
Ready work query:
SELECT id, title, priority FROM agent_tasks t WHERE t.deleted_at IS NULL AND t.status IN ('open','in_progress') AND NOT EXISTS (SELECT 1 FROM blocked_tasks_cache b WHERE b.task_id = t.id) ORDER BY priority ASC, updated_at ASC LIMIT 50;
Claim a task (atomic):
SELECT claim_task(2, 'agent-1');
Link a task to a memory:
INSERT INTO task_memory_links(task_id, memory_id, link_type) VALUES (2, 123, 'outcome');
Optional add-on: conditional_blocks (not implemented yet)
-
This is intentionally deferred until the core workflow feels solid.
-
If you need it now, store a condition in task_links.metadata (e.g., { "os": "windows" } ) and treat it as documentation.
Compaction Log (high value)
Compaction can delete context. Treat every compaction as an important event and record it.
If you're using OpenCode, prefer the OpenCode plugin route for automatic compaction logging.
OpenCode plugin (experimental.session.compacting)
-
Copy plugins/agent-memory-systems-postgres.js to ~/.config/opencode/plugins/
-
Restart OpenCode
It writes local compaction events to:
- ~/.config/opencode/agent-memory-systems-postgres/compaction-events.jsonl
And will also attempt a best-effort Postgres store_memory(...) write (requires pgpass).
Verify
SELECT id, title, relevance_score FROM search_memories('compaction', NULL, NULL, NULL, NULL, 0, 10);
If nothing is inserted, set up .pgpass / pgpass.conf so psql can authenticate without prompting.
Daily Compaction Consolidation
Raw compaction events are noisy. Run a daily consolidation job that turns many compaction events into 1 daily memory.
The consolidation scripts default to the OpenCode plugin event log path and will fall back to Claude Code paths if needed.
-
OpenCode events: ~/.config/opencode/agent-memory-systems-postgres/compaction-events.jsonl
-
Output directory: ~/.config/opencode/agent-memory-systems-postgres/compaction-daily/
Windows run (manual):
powershell.exe -NoProfile -ExecutionPolicy Bypass -File "scripts\consolidate-compactions.ps1"
Linux/macOS run (manual):
bash "scripts/consolidate-compactions.sh"
Scheduling:
-
Windows Task Scheduler: create a daily task that runs the PowerShell command above
-
Linux cron example:
every day at 02:10 UTC
10 2 * * * bash "<skill-dir>/scripts/consolidate-compactions.sh" >/dev/null 2>&1
Appendix: Claude Code compatibility (optional)
This repository also includes Claude Code hook scripts under hooks/ . They are not required for OpenCode usage.
Friction Log (turn pain into tooling)
Whenever something is annoying, brittle, or fails:
-
Store an episodic memory with category friction and tags for the tool/OS/error.
-
If it repeats (2+ times), promote it to procedural memory (importance >= 7) with a checklist.
-
Update this skill doc when the fix becomes a stable rule/workflow (so every agent learns it).
Schema Overview
agent_memories — Main table. Full-text search, trigram indexes, JSONB metadata, soft-delete. memory_links — Graph relationships (references, supersedes, contradicts). working_memory — Ephemeral session context with auto-expire.
Key columns: memory_type , category , tags[] , title , content , content_hash (auto), metadata (JSONB), importance_score , access_count , relevance_decay , search_vector (auto).
Anti-Patterns
Don't Do Instead
Store everything Only store non-obvious solutions
Skip tags Tag comprehensively: tech, error codes, platform
Use to_tsquery directly search_memories() handles this via plainto_tsquery
One type for all data Use correct memory_type per content
Forget importance rating Rate honestly: 9-10 battle-tested, 5-6 partial
Sharp Edges
Issue Severity Mitigation
Chunks lose context Critical Store full problem+solution as one unit
Old tech memories High apply_memory_decay() daily; prune stale
Duplicate memories Medium store_memory() auto-deduplicates by content_hash
No vector search Info pg_trgm provides fuzzy matching; pgvector can be added later
Cross-Platform Notes
-
PostgreSQL 14-18 supported (no partitioning, no GENERATED ALWAYS)
-
pg_trgm is the only required extension (ships with all PG distributions)
-
Linux: psql -U postgres -d agent_memory -f init.sql
-
Windows: Use full path to psql.exe or add PG bin to PATH
-
MCP postgres_query: Works for read operations; DDL requires psql
Maintenance
SELECT apply_memory_decay(); -- daily SELECT prune_stale_memories(180, 3.0, 0); -- monthly DELETE FROM working_memory WHERE expires_at < NOW(); -- daily SELECT * FROM memory_health_check(); -- anytime
Optional: pgvector Semantic Search
If pgvector is installed on your PostgreSQL server, init.sql will:
-
create extension vector (non-fatal if missing)
-
add agent_memories.embedding vector (variable dimension)
-
create search_memories_vector(p_embedding, p_embedding_dim, ...)
Notes:
-
This does NOT generate embeddings. You must populate agent_memories.embedding yourself.
-
Once embeddings exist, you can do nearest-neighbor search:
-- p_embedding is a pgvector literal; pass it from your app. -- Optionally filter by dimension (recommended when using multiple models). SELECT id, title, similarity FROM search_memories_vector('[0.01, 0.02, ...]'::vector, 768, NULL, NULL, NULL, NULL, 0.0, 10);
Note: variable-dimension vectors cannot be indexed with pgvector indexes. This is a tradeoff to support local models with different embedding sizes.
If pgvector is not installed, everything else still works (fts + pg_trgm).
Embedding Ingestion Pipeline
pgvector search only works after you populate agent_memories.embedding .
This skill ships ingestion scripts (relative paths). Run from the skill directory:
-
scripts/ingest-embeddings.ps1
-
scripts/ingest-embeddings.sh
They:
-
find memories with embedding IS NULL
-
call an OpenAI-compatible embeddings endpoint (including Ollama)
-
write vectors into agent_memories.embedding vector
Requirements:
-
pgvector installed + init.sql applied (so agent_memories.embedding exists)
-
.pgpass / pgpass.conf configured (so psql -w can write without prompting)
-
env vars for embedding API:
-
EMBEDDING_PROVIDER (ollama or openai ; default openai )
-
EMBEDDING_API_KEY (required for openai ; optional for ollama )
-
EMBEDDING_API_URL (default depends on provider)
-
EMBEDDING_MODEL (default depends on provider)
-
EMBEDDING_DIMENSIONS (optional; forwarded to the embeddings endpoint when supported)
Windows example:
$env:EMBEDDING_PROVIDER = "ollama" $env:EMBEDDING_MODEL = "nomic-embed-text" powershell.exe -NoProfile -ExecutionPolicy Bypass -File "scripts\ingest-embeddings.ps1" -Limit 25
Linux/macOS example:
export EMBEDDING_API_KEY=... export EMBEDDING_MODEL=text-embedding-3-small bash "scripts/ingest-embeddings.sh"
Scheduling:
-
run daily (or hourly) after you add new memories
-
keep Limit small until you trust it
Related Skills
systematic-debugging , postgres-pro , postgresql-table-design