Synapse AI Agent Orchestration Skill
Synapse is a distributed AI agent system that orchestrates three specialized services with configurable models and role-based workflows.
Agent Roles
| Agent | Role | Capabilities | Default Model |
|---|---|---|---|
| Claude | Orchestrator/Planner | Task planning, architecture design, code generation | claude-sonnet-4.5 |
| Gemini | Analyst/Reviewer | Large context analysis (>1M tokens), code review, security audits | gemini-3-pro-preview |
| Codex | Executor | Command execution, build processes, automated testing | gpt-5.2 |
When to Apply
Use this skill when:
- Breaking down complex tasks into actionable multi-step plans
- Generating implementation code from descriptions or specifications
- Analyzing large codebases or content (especially >200k tokens via Gemini)
- Getting comprehensive code reviews with quality scoring
- Running sandboxed shell commands for builds, tests, or automation
- Executing full plan-analyze-code-review-execute workflows
- Orchestrating parallel or pipeline-based multi-agent workflows
Do NOT use when:
- Simple questions that don't require code generation
- Tasks that only need local file operations
- When Docker/Synapse services are not available
Prerequisites
Docker must be running with Synapse services active.
# Clone Synapse (if not already)
git clone https://github.com/akillness/Synapse.git ~/Synapse
cd ~/Synapse
# Start services
docker compose up -d
# Verify all services are healthy
curl -s http://localhost:8000/health
# Expected: {"status": "healthy", "service": "gateway"}
Installation
Global Installation (Recommended)
# Clone synapse-skill
git clone https://github.com/akillness/synapse-skill.git
# For OpenCode
ln -s $(pwd)/synapse-skill ~/.config/opencode/skills/synapse
# For Claude Code / Other Agents
mkdir -p ~/.agents/skills
ln -s $(pwd)/synapse-skill ~/.agents/skills/synapse
Environment Variables (Optional)
# Custom Synapse location
export SYNAPSE_HOME="$HOME/Synapse"
# Gateway URL (default: http://localhost:8000)
export SYNAPSE_GATEWAY_URL="http://localhost:8000"
For Cursor IDE
Create .cursor/rules/synapse.mdc in your project:
---
description: Synapse AI Agent Orchestration
globs: ["**/*"]
alwaysApply: true
---
# Synapse Skill Integration
When orchestrating multi-agent tasks, use Synapse endpoints:
- Plan: POST http://localhost:8000/api/v1/claude/plan
- Code: POST http://localhost:8000/api/v1/claude/code
- Analyze: POST http://localhost:8000/api/v1/gemini/analyze
- Review: POST http://localhost:8000/api/v1/gemini/review
- Execute: POST http://localhost:8000/api/v1/codex/execute
Use @planner for task breakdown, @reviewer for code review.
For Claude Code (CLAUDE.md)
Add to your ~/.claude/CLAUDE.md:
## Synapse Integration
For multi-agent orchestration, use Synapse skill:
- Health check: curl http://localhost:8000/health
- Ensure Docker containers are running: docker ps | grep synaps
- Use /api/v1/workflow for full pipeline execution
For OpenCode (oh-my-opencode.json)
Add Synapse-optimized agent configuration to ~/.config/opencode/oh-my-opencode.json:
{
"$schema": "https://raw.githubusercontent.com/code-yeongyu/oh-my-opencode/master/assets/oh-my-opencode.schema.json",
"agents": {
"synapse-planner": {
"model": "google/antigravity-claude-sonnet-4-5"
},
"synapse-analyst": {
"model": "google/antigravity-gemini-3-pro-high"
},
"synapse-coder": {
"model": "google/antigravity-claude-sonnet-4-5-thinking"
},
"synapse-reviewer": {
"model": "google/antigravity-gemini-3-pro-high"
},
"synapse-executor": {
"model": "openai/gpt-5.2-codex-high"
}
}
}
OpenCode Model Provider Configuration
Add to ~/.config/opencode/opencode.json for custom Synapse models:
{
"provider": {
"synapse": {
"name": "Synapse Local",
"api": "openai",
"options": {
"baseURL": "http://localhost:8000/api/v1/"
},
"models": {
"claude-planner": {
"name": "Claude Planner (Synapse)",
"limit": { "context": 200000, "output": 64000 }
},
"gemini-analyst": {
"name": "Gemini Analyst (Synapse)",
"limit": { "context": 1000000, "output": 65536 }
},
"codex-executor": {
"name": "Codex Executor (Synapse)",
"limit": { "context": 272000, "output": 128000 }
}
}
}
}
}
OpenCode Skill Integration
The skill is automatically loaded when symlinked to ~/.config/opencode/skills/synapse.
Trigger phrases for OpenCode:
- "orchestrate agents" → Activates Synapse workflow
- "multi-agent workflow" → Starts parallel or pipeline execution
- "run synapse" → Executes full workflow
Model Configuration
Claude Service Models (2026)
| Model | API ID | Best For | Context | Cost (Input/Output) |
|---|---|---|---|---|
claude-opus-4.5 | claude-opus-4-5-20251101 | Complex reasoning, production code, sophisticated agents | 200K | $5/$25 per M |
claude-sonnet-4.5 | claude-sonnet-4-5-20250929 | Task planning, code generation, architecture (Recommended) | 1M | $3/$15 per M |
claude-haiku-4.5 | claude-haiku-4-5-20251201 | Fast responses, simple tasks, 90% of Sonnet performance | 200K | $0.80/$4 per M |
Claude 4.5 Highlights:
- Opus 4.5: State-of-the-art coding (80%+ SWE-bench), best for agents and computer use
- Sonnet 4.5: 1M token context window (5x increase), extended thinking mode
- Haiku 4.5: Cost-effective with 90% of Sonnet's performance
Gemini Service Models (2026)
| Model | Best For | Context | Cost |
|---|---|---|---|
gemini-3-pro-preview | Complex reasoning, coding, agentic tasks (76.2% SWE-bench) | 1M | $2-4/M |
gemini-3-flash | Sub-second latency, speed-critical, distilled from 3 Pro | 1M | Lower |
gemini-2.5-pro | Large context analysis, code review, mature stability | 1M | $1.25/M |
gemini-2.5-flash | Cost-efficient, high-volume ($0.15/M input) | 1M | $0.15/M |
Gemini 3 Highlights:
- 35% better at software engineering tasks vs 2.5 Pro
- "Vibe coding" support for rapid prototyping
- Knowledge cutoff: January 2025
Codex Service Models
| Model | Best For | Context | Cost |
|---|---|---|---|
gpt-5.2 | Software engineering, agentic workflows | 400K | $1.25/$10 |
gpt-5.2-mini | Cost-efficient coding (4x more usage) | 400K | $0.25/$2 |
gpt-5.1-thinking | Ultra-complex reasoning | 400K | Higher |
Agentic Workflow Patterns
Pattern 1: Parallel Specialists
Multiple specialists review code simultaneously:
┌─────────────────────────────────────────────────────┐
│ ORCHESTRATOR │
│ (You / Claude) │
└────────────┬───────────┬───────────┬────────────────┘
│ │ │
┌───────▼───┐ ┌─────▼─────┐ ┌───▼───────┐
│ Security │ │Performance│ │ Simplicity│
│ Reviewer │ │ Reviewer │ │ Reviewer │
│ (Gemini) │ │ (Gemini) │ │ (Gemini) │
└───────────┘ └───────────┘ └───────────┘
Workflow:
# 1. Spawn parallel reviews
curl -X POST http://localhost:8000/api/v1/gemini/review \
-d '{"code": "<code>", "language": "python", "review_type": "security"}'
curl -X POST http://localhost:8000/api/v1/gemini/review \
-d '{"code": "<code>", "language": "python", "review_type": "performance"}'
curl -X POST http://localhost:8000/api/v1/gemini/review \
-d '{"code": "<code>", "language": "python", "review_type": "simplicity"}'
# 2. Aggregate results and synthesize
Pattern 2: Pipeline (Sequential Dependencies)
Each stage depends on the previous:
┌──────────┐ ┌──────────┐ ┌──────────┐ ┌──────────┐ ┌──────────┐
│ Research │───▶│ Plan │───▶│Implement │───▶│ Test │───▶│ Review │
│ (Gemini) │ │ (Claude) │ │ (Claude) │ │ (Codex) │ │ (Gemini) │
└──────────┘ └──────────┘ └──────────┘ └──────────┘ └──────────┘
Workflow:
# Step 1: Research (Gemini analyzes requirements)
RESEARCH=$(curl -X POST http://localhost:8000/api/v1/gemini/analyze \
-d '{"content": "<requirements>", "analysis_type": "documentation"}')
# Step 2: Plan (Claude creates implementation plan)
PLAN=$(curl -X POST http://localhost:8000/api/v1/claude/plan \
-d "{\"task\": \"Implement based on: $RESEARCH\", \"constraints\": [\"python\", \"tests\"]}")
# Step 3: Implement (Claude generates code)
CODE=$(curl -X POST http://localhost:8000/api/v1/claude/code \
-d "{\"description\": \"$PLAN\", \"language\": \"python\"}")
# Step 4: Test (Codex executes tests)
TEST=$(curl -X POST http://localhost:8000/api/v1/codex/execute \
-d '{"command": "python -m pytest tests/", "timeout": 60}')
# Step 5: Review (Gemini reviews final code)
REVIEW=$(curl -X POST http://localhost:8000/api/v1/gemini/review \
-d "{\"code\": \"$CODE\", \"language\": \"python\"}")
Pattern 3: Swarm (Self-Organizing)
Workers grab available tasks from a pool:
┌─────────────┐
│ Task Pool │
│ ┌─┬─┬─┬─┬─┐ │
│ │1│2│3│4│5│ │
│ └─┴─┴─┴─┴─┘ │
└──────┬──────┘
┌───────────────┼───────────────┐
│ │ │
┌─────▼─────┐ ┌─────▼─────┐ ┌─────▼─────┐
│ Worker 1 │ │ Worker 2 │ │ Worker 3 │
│ (Claude) │ │ (Gemini) │ │ (Codex) │
│ claims #1 │ │ claims #2 │ │ claims #3 │
└───────────┘ └───────────┘ └───────────┘
Workflow:
# Create task pool via workflow endpoint
curl -X POST http://localhost:8000/api/v1/workflow \
-d '{
"task": "Review all files in src/",
"mode": "swarm",
"workers": 3,
"constraints": ["parallel", "auto-assign"]
}'
Pattern 4: Full Workflow (End-to-End)
Use the workflow endpoint for automatic orchestration:
curl -X POST http://localhost:8000/api/v1/workflow \
-H "Content-Type: application/json" \
-d '{
"task": "Build a user authentication module",
"constraints": ["FastAPI", "JWT", "PostgreSQL"],
"workflow_type": "pipeline",
"model_config": {
"planner": "claude-sonnet-4",
"analyst": "gemini-2.5-pro",
"coder": "claude-sonnet-4",
"executor": "gpt-5.2"
}
}'
Role-Based Model Assignment
Configuration Schema
{
"roles": {
"planner": {
"service": "claude",
"model": "claude-sonnet-4.5",
"description": "Creates task plans and architecture designs"
},
"analyst": {
"service": "gemini",
"model": "gemini-3-pro-preview",
"description": "Analyzes content and performs code reviews"
},
"coder": {
"service": "claude",
"model": "claude-sonnet-4.5",
"description": "Generates implementation code"
},
"reviewer": {
"service": "gemini",
"model": "gemini-3-pro-preview",
"description": "Performs comprehensive code reviews"
},
"executor": {
"service": "codex",
"model": "gpt-5.2",
"description": "Executes commands and runs tests"
}
}
}
Role Selection by Task Type
| Task Type | Primary Role | Model | Endpoint |
|---|---|---|---|
| Task breakdown | Planner | claude-sonnet-4 | /api/v1/claude/plan |
| Code generation | Coder | claude-sonnet-4 | /api/v1/claude/code |
| Large context analysis | Analyst | gemini-2.5-pro | /api/v1/gemini/analyze |
| Security review | Reviewer | gemini-3-pro-preview | /api/v1/gemini/review |
| Build & test | Executor | gpt-5.2 | /api/v1/codex/execute |
API Reference
Base URL
http://localhost:8000
Endpoints Summary
| Service | Method | Endpoint | Purpose |
|---|---|---|---|
| System | GET | /health | Gateway health check |
| System | GET | /metrics | Pool and load balancer stats |
| Claude | GET | /api/v1/claude/health | Service health |
| Claude | POST | /api/v1/claude/plan | Create task plan |
| Claude | POST | /api/v1/claude/code | Generate code |
| Gemini | GET | /api/v1/gemini/health | Service health |
| Gemini | POST | /api/v1/gemini/analyze | Analyze content |
| Gemini | POST | /api/v1/gemini/review | Review code |
| Codex | GET | /api/v1/codex/health | Service health |
| Codex | POST | /api/v1/codex/execute | Execute command |
| Workflow | POST | /api/v1/workflow | Full pipeline |
Quick Reference
| Use Case | Service | Endpoint | Key Fields |
|---|---|---|---|
| Break down task | Claude | /api/v1/claude/plan | task, constraints, model |
| Generate code | Claude | /api/v1/claude/code | description, language, model |
| Analyze codebase | Gemini | /api/v1/gemini/analyze | content, analysis_type, model |
| Code review | Gemini | /api/v1/gemini/review | code, language, review_type, model |
| Run command | Codex | /api/v1/codex/execute | command, timeout, model |
| Full pipeline | Workflow | /api/v1/workflow | task, constraints, workflow_type, model_config |
Running Tasks
1. Health Check (Always Start Here)
# Check all services
curl -s http://localhost:8000/api/v1/claude/health
curl -s http://localhost:8000/api/v1/gemini/health
curl -s http://localhost:8000/api/v1/codex/health
Expected response:
{"status": "SERVING", "version": "1.0.0", "uptime_seconds": 123}
2. Create a Plan (Claude)
curl -X POST http://localhost:8000/api/v1/claude/plan \
-H "Content-Type: application/json" \
-d '{
"task": "Build a REST API with authentication",
"constraints": ["use FastAPI", "JWT tokens", "PostgreSQL"],
"model": "claude-sonnet-4"
}'
3. Generate Code (Claude)
curl -X POST http://localhost:8000/api/v1/claude/code \
-H "Content-Type: application/json" \
-d '{
"description": "A function to validate email addresses using regex",
"language": "python",
"model": "claude-sonnet-4"
}'
4. Analyze Content (Gemini)
curl -X POST http://localhost:8000/api/v1/gemini/analyze \
-H "Content-Type: application/json" \
-d '{
"content": "<large_codebase_content>",
"analysis_type": "code",
"model": "gemini-2.5-pro"
}'
5. Review Code (Gemini)
curl -X POST http://localhost:8000/api/v1/gemini/review \
-H "Content-Type: application/json" \
-d '{
"code": "def add(a, b):\n return a + b",
"language": "python",
"review_type": "comprehensive",
"model": "gemini-3-pro-preview"
}'
6. Execute Command (Codex)
curl -X POST http://localhost:8000/api/v1/codex/execute \
-H "Content-Type: application/json" \
-d '{
"command": "python -m pytest tests/",
"working_dir": "/tmp",
"timeout": 60,
"model": "gpt-5.2"
}'
Allowed Commands: echo, ls, pwd, date, cat, head, tail, wc, grep, find, python, pip, npm, node, git, make
Common Workflows
Code Development Workflow
# 1. Create plan
curl -X POST http://localhost:8000/api/v1/claude/plan \
-H "Content-Type: application/json" \
-d '{"task": "Implement user authentication", "model": "claude-sonnet-4"}'
# 2. Generate code
curl -X POST http://localhost:8000/api/v1/claude/code \
-H "Content-Type: application/json" \
-d '{"description": "JWT authentication middleware", "language": "python"}'
# 3. Review code
curl -X POST http://localhost:8000/api/v1/gemini/review \
-H "Content-Type: application/json" \
-d '{"code": "<generated_code>", "language": "python", "model": "gemini-3-pro-preview"}'
# 4. Run tests
curl -X POST http://localhost:8000/api/v1/codex/execute \
-H "Content-Type: application/json" \
-d '{"command": "python -m pytest tests/", "timeout": 60}'
Code Analysis Workflow
# 1. Read file content
FILE_CONTENT=$(cat myfile.py | jq -Rs .)
# 2. Analyze with large context model
curl -X POST http://localhost:8000/api/v1/gemini/analyze \
-H "Content-Type: application/json" \
-d "{\"content\": $FILE_CONTENT, \"analysis_type\": \"code\", \"model\": \"gemini-2.5-pro\"}"
# 3. Review with detailed model
curl -X POST http://localhost:8000/api/v1/gemini/review \
-H "Content-Type: application/json" \
-d "{\"code\": $FILE_CONTENT, \"language\": \"python\", \"model\": \"gemini-3-pro-preview\"}"
Error Handling
Service Unavailable (503)
{"detail": "Service not ready"}
Solution: Check if Docker containers are running:
docker ps | grep synaps
Connection Refused
Solution: Start Synapse services:
cd ~/Synapse # or $SYNAPSE_HOME
docker compose up -d
Command Not Allowed (Codex)
{"success": false, "stderr": "Command not allowed"}
Solution: Use only allowed commands (echo, ls, pwd, python, etc.)
Troubleshooting
Check Service Status
# All containers running?
docker ps --format "table {{.Names}}\t{{.Status}}" | grep synaps
# Gateway health
curl -s http://localhost:8000/health
# Individual service health
curl -s http://localhost:8000/api/v1/claude/health
curl -s http://localhost:8000/api/v1/gemini/health
curl -s http://localhost:8000/api/v1/codex/health
Restart Services
cd ~/Synapse # or $SYNAPSE_HOME
docker compose down
docker compose up -d
View Logs
docker logs synaps-gateway
docker logs synaps-claude
docker logs synaps-gemini
docker logs synaps-codex
Service Ports
| Service | Port | Protocol |
|---|---|---|
| Gateway | 8000 | HTTP/REST |
| Claude | 5011 | gRPC |
| Gemini | 5012 | gRPC |
| Codex | 5013 | gRPC |
| Prometheus | 9090 | HTTP |
| Grafana | 3000 | HTTP |
Integration with AI Agents
This skill can be used by any AI coding agent that supports HTTP requests:
- OpenCode/Claude Code: Use bash tool with curl commands
- Cursor: Use terminal or HTTP client extensions
- GitHub Copilot: Use inline curl in terminal
- Windsurf: Use built-in terminal
Example Agent Integration
User: "Create a plan to build a web scraper"
Agent uses Synapse skill:
1. curl POST /api/v1/claude/plan with task and model config
2. Parse response and present steps to user
3. Offer to generate code for each step
4. Execute tests via Codex
5. Review final code via Gemini
Swarm Mode Integration
For complex multi-file tasks, use swarm orchestration:
User: "Review all authentication files for security issues"
Agent orchestrates:
1. Spawn parallel Gemini reviewers for each file
2. Aggregate findings
3. Generate summary report
4. Propose fixes via Claude
Current Status (v2.3.0)
Service Health (Verified 2026-01-28)
| Service | Endpoint | Status | Response |
|---|---|---|---|
| Gateway | /health | ✅ healthy | {"status": "healthy"} |
| Claude | /api/v1/claude/plan | ✅ success | 5-step plan generated |
| Gemini | /api/v1/gemini/analyze | ✅ success | Code analysis complete |
| Gemini | /api/v1/gemini/review | ✅ success | Score 95/100 |
| Codex | /api/v1/codex/execute | ✅ success | Command executed |
Connection Pool Metrics
Claude Pool Gemini Pool Codex Pool
┌────────────┐ ┌────────────┐ ┌────────────┐
│ Size: 2 │ │ Size: 2 │ │ Size: 2 │
│ Avail: 2 │ │ Avail: 2 │ │ Avail: 2 │
│ Acquired:3 │ │ Acquired:4 │ │ Acquired:3 │
│ Released:3 │ │ Released:4 │ │ Released:3 │
└────────────┘ └────────────┘ └────────────┘
Load Balancer Status: All endpoints healthy (1/1 each)
Avg Response Time (Claude): 21.47ms
Architecture Overview
┌─────────────────────────────────────────────────────────────────────────────┐
│ SYNAPSE WORKFLOW ARCHITECTURE │
└─────────────────────────────────────────────────────────────────────────────┘
┌──────────────┐
│ Gateway │ :8000
│ (REST) │
└──────┬───────┘
│
┌──────────────────────┼──────────────────────┐
│ │ │
▼ ▼ ▼
┌────────────────┐ ┌────────────────┐ ┌────────────────┐
│ Claude │ │ Gemini │ │ Codex │
│ :5011 │ │ :5012 │ │ :5013 │
│ (gRPC) │ │ (gRPC) │ │ (gRPC) │
└────────────────┘ └────────────────┘ └────────────────┘
│ │ │
▼ ▼ ▼
┌────────────┐ ┌────────────┐ ┌────────────┐
│ /plan │ │ /analyze │ │ /execute │
│ /code │ │ /review │ │ │
└────────────┘ └────────────┘ └────────────┘
Improvement Roadmap
1. Unified Workflow Endpoint Enhancement
Current: Individual service calls required
curl /api/v1/claude/plan → curl /api/v1/claude/code → curl /api/v1/gemini/review → curl /api/v1/codex/execute
Planned: Single workflow call with workflow_type parameter
curl /api/v1/workflow -d '{"task": "...", "workflow_type": "pipeline|parallel|swarm"}'
2. MCP Tool Direct Integration
Current: curl-based access via Bash tool
Bash → curl → Gateway → Service
Planned: Native MCP server tools
mcp__synapse__plan → Gateway → Claude
mcp__synapse__analyze → Gateway → Gemini
mcp__synapse__execute → Gateway → Codex
3. Enhanced Error Handling
Current: Simple error response
{"detail": "Service not ready"}
Planned: Rich error with retry guidance
{
"error": "Service not ready",
"retry_after": 5,
"fallback_available": true,
"fallback_service": "gemini-2.5-flash"
}
4. SSE Streaming Response Support
Current: Full response wait
response = await service.plan(request) # Waits for completion
Planned: Real-time progress streaming
async for chunk in service.plan_stream(request):
yield chunk # Real-time progress updates
5. Redis Caching Layer
Planned Architecture:
Request
│
▼
┌─────────────┐ Cache Hit ┌─────────────┐
│ Gateway │ ─────────────────▶ │ Redis │
└──────┬──────┘ └─────────────┘
│ Cache Miss
▼
┌─────────────┐
│ Service │
└─────────────┘
6. Quick Commands (Planned)
# Shorthand commands for common operations
synapse plan "Build REST API" # → Claude Plan
synapse review "def foo(): ..." # → Gemini Review
synapse run "pytest tests/" # → Codex Execute
synapse flow "Create feature X" # → Full Pipeline
7. Auto Service Detection
Planned Features:
- Docker container status check on skill activation
- Automatic unhealthy service restart prompt
- Service dependency validation