MCP Engineering

# MCP Engineering — Complete Model Context Protocol System

Safety Notice

This item is sourced from the public archived skills repository. Treat as untrusted until reviewed.

Copy this and send it to your AI assistant to learn

Install skill "MCP Engineering" with this command: npx skills add 1kalin/afrexai-mcp-engineering

MCP Engineering — Complete Model Context Protocol System

Build, integrate, secure, and scale MCP servers and clients. From first server to production multi-tool architecture.

When to Use

  • Building an MCP server (any language)
  • Integrating MCP tools into an AI agent
  • Debugging MCP connection/auth issues
  • Designing multi-server architectures
  • Securing MCP endpoints for production
  • Evaluating which MCP servers to use

Phase 1: MCP Fundamentals

What MCP Is

Model Context Protocol = standardized way for AI agents to call external tools. Think of it as "USB for AI" — one protocol, any tool.

Architecture

Agent (Client) ←→ MCP Transport ←→ MCP Server ←→ External Service
                   (stdio/HTTP)      (your code)    (API, DB, file system)

Core Concepts

ConceptWhat It DoesExample
ServerExposes tools, resources, promptsA server wrapping the GitHub API
ClientDiscovers and calls server capabilitiesOpenClaw, Claude Desktop, Cursor
ToolA callable function with typed paramscreate_issue(title, body, labels)
ResourceRead-only data the agent can accessfile://workspace/config.json
PromptReusable prompt templatessummarize_pr(pr_url)
TransportHow client↔server communicatestdio (local) or HTTP+SSE (remote)

Transport Decision

FactorstdioHTTP/SSEStreamable HTTP
Setup complexityLowMediumMedium
Multi-clientNoYesYes
Remote accessNoYesYes
StreamingVia stdioSSENative
Auth neededNo (local)YesYes
Best forLocal dev, single agentProduction, sharedModern production

Rule: Start with stdio for development. Move to HTTP for production or multi-agent.


Phase 2: Building Your First MCP Server

Server Brief YAML

server_name: "[service]-mcp"
description: "[What this server does in one sentence]"
transport: stdio | http
tools:
  - name: "[verb_noun]"
    description: "[What it does — be specific for LLM tool selection]"
    params:
      - name: "[param]"
        type: "string | number | boolean | object | array"
        required: true | false
        description: "[What this param controls]"
    returns: "[What the tool returns]"
    error_cases:
      - "[When/how it fails]"
resources:
  - uri: "[protocol://path]"
    description: "[What data this exposes]"
external_dependencies:
  - "[API/service this wraps]"
auth_required: true | false
auth_method: "api_key | oauth2 | none"

TypeScript Server Template (stdio)

// server.ts — minimal MCP server
import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StdioServerTransport } from "@modelcontextprotocol/sdk/server/stdio.js";
import { z } from "zod";

const server = new McpServer({
  name: "my-service",
  version: "1.0.0",
});

// Define a tool
server.tool(
  "get_item",                          // tool name (verb_noun)
  "Fetch an item by ID",               // description (LLM reads this)
  { id: z.string().describe("Item ID") }, // params with descriptions
  async ({ id }) => {
    try {
      const result = await fetchItem(id);
      return {
        content: [{ type: "text", text: JSON.stringify(result, null, 2) }],
      };
    } catch (error) {
      return {
        content: [{ type: "text", text: `Error: ${error.message}` }],
        isError: true,
      };
    }
  }
);

// Define a resource
server.resource(
  "config",
  "config://app",
  async (uri) => ({
    contents: [{ uri: uri.href, mimeType: "application/json", text: JSON.stringify(config) }],
  })
);

// Start
const transport = new StdioServerTransport();
await server.connect(transport);

Python Server Template (stdio)

# server.py — minimal MCP server
from mcp.server import Server
from mcp.server.stdio import stdio_server
from mcp.types import Tool, TextContent
import json

server = Server("my-service")

@server.list_tools()
async def list_tools():
    return [
        Tool(
            name="get_item",
            description="Fetch an item by ID",
            inputSchema={
                "type": "object",
                "properties": {
                    "id": {"type": "string", "description": "Item ID"}
                },
                "required": ["id"]
            }
        )
    ]

@server.call_tool()
async def call_tool(name: str, arguments: dict):
    if name == "get_item":
        result = await fetch_item(arguments["id"])
        return [TextContent(type="text", text=json.dumps(result, indent=2))]
    raise ValueError(f"Unknown tool: {name}")

async def main():
    async with stdio_server() as (read, write):
        await server.run(read, write, server.create_initialization_options())

if __name__ == "__main__":
    import asyncio
    asyncio.run(main())

Tool Design Rules

  1. Verb-noun naming: create_issue, search_docs, update_config — never issue or doStuff
  2. Descriptions are critical: The LLM picks tools based on descriptions. Be specific. Include when NOT to use.
  3. Granular over god-tools: search_issues + get_issue + create_issue beats manage_issues
  4. Return structured data: JSON over prose. Let the LLM format for the user.
  5. Error messages for LLMs: Include what went wrong AND what to try next
  6. Idempotent where possible: create_or_update > create (prevents duplicates from retries)
  7. Limit output size: Paginate or truncate. A 10MB response kills the context window.
  8. Include examples in descriptions: "Search issues. Example: search_issues(query='bug label:critical')"

Tool Description Quality Checklist

  • Says what the tool DOES (not just the name restated)
  • Mentions when to use vs. when NOT to use
  • Each param has a description with format hints
  • Return format is documented
  • Edge cases mentioned (empty results, not found, etc.)

Phase 3: HTTP Transport & Production Server

HTTP Server Template (TypeScript)

import { McpServer } from "@modelcontextprotocol/sdk/server/mcp.js";
import { StreamableHTTPServerTransport } from "@modelcontextprotocol/sdk/server/streamableHttp.js";
import express from "express";

const app = express();
app.use(express.json());

const server = new McpServer({ name: "my-service", version: "1.0.0" });
// ... register tools ...

app.post("/mcp", async (req, res) => {
  const transport = new StreamableHTTPServerTransport("/mcp", res);
  await server.connect(transport);
  await transport.handleRequest(req, res);
});

app.listen(3001, () => console.log("MCP server on :3001"));

Auth Patterns

API Key (simplest)

// Middleware
function authMiddleware(req, res, next) {
  const key = req.headers["x-api-key"] || req.headers.authorization?.replace("Bearer ", "");
  if (!key || !validKeys.has(key)) {
    return res.status(401).json({ error: "Invalid API key" });
  }
  req.userId = keyToUser.get(key);
  next();
}

OAuth 2.0 (for user-scoped access)

# MCP OAuth flow
1. Client requests tool → server returns 401 with auth URL
2. User completes OAuth in browser → gets access token
3. Client stores token, includes in subsequent requests
4. Server validates token, calls external API on user's behalf

Production Checklist

  • Rate limiting per client/key
  • Request validation (schema check before execution)
  • Structured logging (request ID, tool name, latency, status)
  • Health check endpoint (/health)
  • Graceful shutdown (finish in-flight requests)
  • Timeout on external calls (don't let tools hang forever)
  • Output size limits (truncate large responses)
  • Error categorization (4xx client vs 5xx server)
  • CORS if browser clients connect
  • TLS in production (always HTTPS)

Phase 4: Client Integration

OpenClaw Configuration

# In openclaw config — stdio server
mcpServers:
  my-service:
    command: "node"
    args: ["path/to/server.js"]
    env:
      API_KEY: "{{env.MY_SERVICE_API_KEY}}"
# HTTP server
mcpServers:
  my-service:
    url: "https://mcp.myservice.com/mcp"
    headers:
      Authorization: "Bearer {{env.MY_SERVICE_TOKEN}}"

Claude Desktop Configuration

{
  "mcpServers": {
    "my-service": {
      "command": "node",
      "args": ["/path/to/server.js"],
      "env": { "API_KEY": "your-key" }
    }
  }
}

Client-Side Tool Selection

When multiple MCP servers are connected, the agent sees ALL tools. Help the agent pick correctly:

  1. Unique tool names: Prefix if needed (github_search vs jira_search)
  2. Clear descriptions: Disambiguate similar tools across servers
  3. Don't overload: 20-30 tools max across all servers. Beyond that, agents get confused.

Multi-Server Architecture

Agent
├── github-mcp (code: create_pr, search_code, list_issues)
├── slack-mcp (comms: send_message, search_messages)
├── postgres-mcp (data: query, list_tables)
└── internal-mcp (business: get_customer, update_pipeline)

Principle: One server per domain. Don't build a mega-server.


Phase 5: Testing MCP Servers

Test Pyramid

        /  E2E  \        Agent actually uses the tool
       / Integration \    Tool calls real API (sandbox)
      /    Unit       \   Business logic without MCP layer

Unit Test Pattern

// Test the tool handler directly, no MCP transport
describe("get_item", () => {
  it("returns item when found", async () => {
    mockDb.findById.mockResolvedValue({ id: "123", name: "Test" });
    const result = await getItemHandler({ id: "123" });
    expect(result.content[0].text).toContain("Test");
  });

  it("returns error for missing item", async () => {
    mockDb.findById.mockResolvedValue(null);
    const result = await getItemHandler({ id: "missing" });
    expect(result.isError).toBe(true);
  });

  it("handles API timeout gracefully", async () => {
    mockDb.findById.mockRejectedValue(new Error("timeout"));
    const result = await getItemHandler({ id: "123" });
    expect(result.isError).toBe(true);
    expect(result.content[0].text).toContain("try again");
  });
});

Integration Test with MCP Inspector

# Use the MCP Inspector to manually test
npx @modelcontextprotocol/inspector node server.js

# Or use mcporter for CLI testing
mcporter call my-service.get_item id=123
mcporter list my-service --schema  # verify tool schemas

Test Checklist Per Tool

  • Happy path returns expected format
  • Missing required params returns clear error
  • Invalid param types return clear error
  • Not-found cases handled (don't throw, return error content)
  • Rate limit / quota exceeded handled
  • Auth failure handled (expired token, invalid key)
  • Large response truncated appropriately
  • Timeout handled (external API slow)
  • Concurrent calls don't interfere

Phase 6: Common MCP Server Patterns

1. API Wrapper (most common)

Wrap an existing REST/GraphQL API as MCP tools.

External API → MCP Server → Agent

Key decisions:

  • Map 1 API endpoint → 1 MCP tool (usually)
  • Simplify params (agent doesn't need every API option)
  • Aggregate related calls (e.g., get user + get user's repos = 1 tool)
  • Cache where safe (reduce API calls)

2. Database Query

Database → MCP Server → Agent

Safety rules:

  • Read-only by default. Write tools require explicit opt-in.
  • Parameterized queries only. NEVER interpolate agent input into SQL.
  • Row limit on all queries (agent can ask for more if needed).
  • Schema as a resource (let agent discover tables/columns).

3. File System

File System → MCP Server → Agent

Safety rules:

  • Sandbox to specific directories. Never allow ../ traversal.
  • Read-only by default. Write requires allowlist.
  • Size limits on reads. Don't send 1GB files through MCP.

4. Multi-Step Workflow

Some tools need to orchestrate multiple steps:

server.tool("deploy_service", "Build, test, and deploy a service", {
  service: z.string(),
  environment: z.enum(["staging", "production"]),
}, async ({ service, environment }) => {
  // Step 1: Build
  const buildResult = await build(service);
  if (!buildResult.success) return error(`Build failed: ${buildResult.error}`);

  // Step 2: Test
  const testResult = await runTests(service);
  if (!testResult.success) return error(`Tests failed: ${testResult.summary}`);

  // Step 3: Deploy (only if build + tests pass)
  if (environment === "production") {
    // Extra safety: require confirmation resource
    return {
      content: [{
        type: "text",
        text: `Ready to deploy ${service} to production. Tests: ${testResult.passed}/${testResult.total} passed. Call confirm_deploy to proceed.`
      }]
    };
  }
  const deployResult = await deploy(service, environment);
  return success(`Deployed ${service} to ${environment}: ${deployResult.url}`);
});

5. Aggregator Server

Combine multiple data sources into unified tools:

GitHub + Jira + PagerDuty → DevOps MCP Server → Agent

One get_service_status tool that queries all three and returns a unified view.


Phase 7: Security & Hardening

Threat Model

ThreatRiskMitigation
Prompt injection via tool outputAgent executes malicious instructions in API responseSanitize output, strip HTML/scripts
Excessive permissionsTool has write access it shouldn'tPrinciple of least privilege per tool
Data exfiltrationAgent sends sensitive data to wrong toolTool allowlists, audit logging
Denial of serviceAgent calls tool in infinite loopRate limiting, circuit breakers
Credential leakageAPI keys in tool responsesStrip sensitive fields from output
SSRFAgent provides URL that hits internal networkURL allowlisting, no private IPs

Security Checklist

  • Every tool has minimum required permissions
  • Write operations require explicit confirmation or are behind feature flags
  • API keys/secrets NEVER appear in tool responses
  • Output sanitized (no HTML, no executable content)
  • Rate limits per tool AND per client
  • Audit log: who called what tool, when, with what params
  • Input validation before any external call
  • URL parameters validated against allowlist (prevent SSRF)
  • Timeout on every external call (max 30s default)
  • Circuit breaker: disable tool if error rate > 50% for 5 min

Dangerous Tool Patterns (Avoid)

❌ server.tool("execute_sql", ..., async ({ query }) => db.raw(query))
❌ server.tool("run_command", ..., async ({ cmd }) => exec(cmd))
❌ server.tool("fetch_url", ..., async ({ url }) => fetch(url))  // SSRF
❌ server.tool("write_file", ..., async ({ path, content }) => fs.writeFile(path, content))

Safe Alternatives

✅ Parameterized queries with allowlisted tables
✅ Predefined commands with argument validation
✅ URL allowlist + no private IP ranges
✅ Write to specific directory + filename validation

Phase 8: Debugging & Troubleshooting

Common Issues

SymptomLikely CauseFix
Tool not appearing in agentSchema error / server not connectedCheck mcporter list or client logs
"Connection refused"Server not running or wrong portVerify process, check port
Tool times outExternal API slow or hangingAdd timeout, check API health
"Invalid params"Schema mismatch between client/serverVerify schema with --schema flag
Agent picks wrong toolAmbiguous descriptionsRewrite descriptions, add "Use this when..."
Agent calls tool in loopTool returning confusing errorReturn clearer error with "do NOT retry"
Large response crashesNo output truncationAdd pagination or character limit
Auth errors intermittentToken expiryImplement token refresh

Debug Workflow

  1. Verify server starts: node server.js — does it start without errors?
  2. List tools: mcporter list my-server --schema — are all tools registered?
  3. Call directly: mcporter call my-server.tool_name param=value — does it return expected output?
  4. Check client config: Is the server path/URL correct? Are env vars set?
  5. Read client logs: Most clients log MCP connection errors
  6. Test with Inspector: npx @modelcontextprotocol/inspector for interactive debugging

Logging Template

server.tool("my_tool", description, schema, async (params) => {
  const requestId = crypto.randomUUID().slice(0, 8);
  console.error(`[${requestId}] my_tool called:`, JSON.stringify(params));
  const start = Date.now();
  try {
    const result = await doWork(params);
    console.error(`[${requestId}] my_tool success: ${Date.now() - start}ms`);
    return success(result);
  } catch (error) {
    console.error(`[${requestId}] my_tool error: ${error.message} (${Date.now() - start}ms)`);
    return errorResponse(error.message);
  }
});

Note: Use console.error for logs in stdio transport (stdout is reserved for MCP protocol).


Phase 9: MCP Server Selection Guide

Evaluating Existing MCP Servers

Score 0-5 per dimension:

DimensionWhat to Check
MaintainedLast commit < 3 months? Issues addressed? Version > 1.0?
SecureNo raw SQL/exec? Auth implemented? Input validated?
Well-typedFull JSON Schema for all tools? Descriptions useful?
TestedHas tests? CI passing?
DocumentedSetup instructions? Tool descriptions? Examples?
LightweightMinimal dependencies? Fast startup?

Score < 15/30: Build your own. Score 15-24: Use with caution. Score 25+: Good to use.

Popular MCP Server Categories

CategoryUse CaseExamples
CodeGitHub, GitLab, code searchgithub-mcp, gitlab-mcp
DataPostgreSQL, SQLite, Snowflakepostgres-mcp, sqlite-mcp
CommsSlack, Discord, emailslack-mcp, gmail-mcp
DocsNotion, Confluence, Google Docsnotion-mcp, gdocs-mcp
DevOpsAWS, GCP, Kubernetes, Terraformaws-mcp, k8s-mcp
SearchBrave, Google, vector storesbrave-search, rag-mcp
FilesLocal FS, S3, Google Drivefilesystem-mcp, s3-mcp
CRMHubSpot, Salesforcehubspot-mcp, sfdc-mcp

Phase 10: Architecture Patterns

Single Agent + Multiple Servers

Agent ──┬── github-mcp
        ├── slack-mcp
        ├── postgres-mcp
        └── custom-mcp

Best for: Most use cases. Simple, effective.

Gateway Pattern

Agent ── MCP Gateway ──┬── server-1
                       ├── server-2
                       └── server-3

Gateway handles: auth, rate limiting, logging, routing. Best for: Enterprise, multi-tenant, compliance requirements.

Agent-per-Domain

Orchestrator Agent
├── Code Agent (github-mcp, gitlab-mcp)
├── Data Agent (postgres-mcp, analytics-mcp)
└── Comms Agent (slack-mcp, email-mcp)

Best for: Complex workflows, specialized agents.

Tool Count Guidelines

Total ToolsRecommendation
1-10Great. Agent handles well.
10-20Good. Ensure distinct descriptions.
20-30Caution. Group by server, review descriptions.
30-50Risk. Consider agent-per-domain pattern.
50+Dangerous. Agent WILL pick wrong tools. Split or use gateway.

Phase 11: Publishing MCP Servers

Package Structure

my-mcp-server/
├── src/
│   ├── server.ts        # MCP server entry
│   ├── tools/           # Tool handlers
│   │   ├── search.ts
│   │   └── create.ts
│   ├── auth.ts          # Auth middleware
│   └── config.ts        # Configuration
├── tests/
│   ├── tools.test.ts
│   └── integration.test.ts
├── package.json
├── tsconfig.json
├── README.md            # Setup + tool docs
└── LICENSE

README Template for MCP Servers

# [Service] MCP Server

[One sentence: what this enables]

## Quick Start
[3 steps max to get running]

## Tools
| Tool | Description | Params |
|------|-------------|--------|
[Table of all tools]

## Configuration
[Env vars, auth setup]

## Examples
[2-3 real usage examples with agent conversation]

npm Publishing

# package.json
{
  "name": "@myorg/service-mcp",
  "version": "1.0.0",
  "bin": { "service-mcp": "./dist/server.js" },
  "files": ["dist"],
  "keywords": ["mcp", "model-context-protocol", "ai-tools"]
}

npm publish

Quality Rubric (0-100)

DimensionWeightWhat to Score
Tool design20%Names, descriptions, granularity, params
Security20%Auth, input validation, output sanitization, least privilege
Reliability15%Error handling, timeouts, circuit breakers
Testing15%Unit + integration coverage, edge cases
Documentation10%Setup, tool docs, examples
Performance10%Response time, output size, caching
Maintainability10%Code structure, types, logging

Score 0-40: Not production ready. 40-70: Usable with caveats. 70-90: Solid. 90+: Excellent.


Common Mistakes

MistakeFix
God-tool that does everythingSplit into focused tools
Vague tool descriptionsWrite descriptions as if explaining to a new hire
No error handlingEvery external call wrapped in try/catch
Returning raw API responsesShape output for agent consumption
No rate limitingAdd per-tool and per-client limits
Ignoring output sizePaginate or truncate responses
Hardcoded credentialsUse env vars or secret manager
No loggingCan't debug what you can't see
Testing only happy pathTest errors, timeouts, edge cases
Building before checkingSearch for existing MCP server first

Natural Language Commands

  • "Build an MCP server for [service]" → Use Phase 2 templates
  • "Add a tool to my MCP server" → Follow tool design rules
  • "Secure my MCP server" → Phase 7 checklist
  • "Debug MCP connection issue" → Phase 8 workflow
  • "Evaluate this MCP server" → Phase 9 scoring
  • "Design multi-server architecture" → Phase 10 patterns
  • "Publish my MCP server" → Phase 11 structure
  • "Convert REST API to MCP" → Phase 6 Pattern 1
  • "Add auth to my MCP server" → Phase 3 auth patterns
  • "Test my MCP server" → Phase 5 checklist
  • "How many tools is too many?" → Phase 10 tool count table
  • "Review my tool descriptions" → Phase 2 quality checklist

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

General

image-gen

Generate AI images from text prompts. Triggers on: "生成图片", "画一张", "AI图", "generate image", "配图", "create picture", "draw", "visualize", "generate an image".

Archived SourceRecently Updated
General

explainer

Create explainer videos with narration and AI-generated visuals. Triggers on: "解说视频", "explainer video", "explain this as a video", "tutorial video", "introduce X (video)", "解释一下XX(视频形式)".

Archived SourceRecently Updated
General

asr

Transcribe audio files to text using local speech recognition. Triggers on: "转录", "transcribe", "语音转文字", "ASR", "识别音频", "把这段音频转成文字".

Archived SourceRecently Updated