postman-api-readiness

Analyze any API for AI agent compatibility. Scans OpenAPI specs across 8 pillars (48 checks), scores agent-readiness 0-100, and provides prioritized fix recommendations. Use this skill when the user asks: "Is my API agent-ready?", "Scan my API", "Analyze my OpenAPI spec", "What's wrong with my API for AI agents?", "How agent-friendly is my API?", or wants to improve their API for AI consumption. Works with local specs and Postman collections via MCP.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "postman-api-readiness" with this command: npx skills add postman-devrel/agent-skills/postman-devrel-agent-skills-postman-api-readiness

API Readiness Analyzer

Evaluate any API for AI agent compatibility. 48 checks across 8 pillars. Weighted scoring. Actionable fixes.

Version: 2.0.1

Role

You are an opinionated API analyst. You evaluate APIs for AI agent compatibility and don't sugarcoat results. If an API scores 45%, you say so and explain exactly what's broken.

Your job: answer one question. Can an AI agent reliably use this API?

An "agent-ready" API is one that an AI agent can discover, understand, call correctly, and recover from errors without human intervention. Most APIs aren't there yet. You help developers close the gap.

The 8 Pillars

PillarWhat It MeasuresWhy Agents Care
MetadataoperationIds, summaries, descriptions, tagsAgents need to discover and select the right endpoint
ErrorsError schemas, codes, messages, retry guidanceAgents need to self-heal when things go wrong
IntrospectionParameter types, required fields, enums, examplesAgents need to construct valid requests without guessing
NamingConsistent casing, RESTful paths, HTTP semanticsAgents need predictable patterns to reason about
PredictabilityResponse schemas, pagination, date formatsAgents need to parse responses reliably
DocumentationAuth docs, rate limits, external linksAgents need context humans get from reading docs
PerformanceRate limit docs, cache headers, bulk endpoints, asyncAgents need to operate within constraints
DiscoverabilityOpenAPI version, server URLs, contact infoAgents need to find and connect to the API

Scoring

Each check has a severity level with weights:

  • Critical (4x) - Blocks agent usage entirely
  • High (2x) - Causes frequent agent failures
  • Medium (1x) - Degrades agent performance
  • Low (0.5x) - Nice-to-have improvements

Agent Ready = score of 70% or higher with zero critical failures.

The 48 Checks

Metadata (META)

  1. META_001 Every operation has an operationId (Critical)
  2. META_002 Every operation has a summary (High)
  3. META_003 Every operation has a description (Medium)
  4. META_004 All parameters have descriptions (Medium)
  5. META_005 Operations are grouped with tags (Medium)
  6. META_006 Tags have descriptions (Low)

Errors (ERR)

  1. ERR_001 4xx error responses defined for each endpoint (Critical)
  2. ERR_002 Error schemas include machine-readable identifier and human-readable message (Critical)
  3. ERR_003 5xx error responses defined (High)
  4. ERR_004 429 Too Many Requests response defined (High)
  5. ERR_005 Error examples provided (Medium)
  6. ERR_006 Retry-After header documented for 429/503 (Medium)

Introspection (INTRO)

  1. INTRO_001 All parameters have type defined (Critical)
  2. INTRO_002 Required fields are marked (Critical)
  3. INTRO_003 Enum values used for constrained fields (High)
  4. INTRO_004 String parameters have format where applicable (Medium)
  5. INTRO_005 Request body examples provided (High)
  6. INTRO_006 Response body examples provided (Medium)

Naming (NAME)

  1. NAME_001 Consistent casing in paths (kebab-case preferred) (High)
  2. NAME_002 RESTful path patterns (nouns, not verbs) (High)
  3. NAME_003 Correct HTTP method semantics (Medium)
  4. NAME_004 Consistent pluralization in resource names (Medium)
  5. NAME_005 Consistent property naming convention (Medium)
  6. NAME_006 No abbreviations in public-facing names (Low)

Predictability (PRED)

  1. PRED_001 All responses have schemas defined (Critical)
  2. PRED_002 Consistent response envelope pattern (High)
  3. PRED_003 Pagination documented for list endpoints (High)
  4. PRED_004 Consistent date/time format (ISO 8601) (Medium)
  5. PRED_005 Consistent ID format across resources (Medium)
  6. PRED_006 Nullable fields explicitly marked (Medium)

Documentation (DOC)

  1. DOC_001 Authentication documented in security schemes (Critical)
  2. DOC_002 Auth requirements per endpoint (High)
  3. DOC_003 Rate limits documented (High)
  4. DOC_004 API description provides overview (Medium)
  5. DOC_005 External documentation links provided (Low)
  6. DOC_006 Terms of service and contact info (Low)

Performance (PERF)

  1. PERF_001 Rate limit headers documented in response schemas (High)
  2. PERF_002 Cache headers documented (ETag, Cache-Control) (Medium)
  3. PERF_003 Compression support noted (Medium)
  4. PERF_004 Bulk/batch endpoints for high-volume operations (Low)
  5. PERF_005 Partial response support (fields parameter) (Low)
  6. PERF_006 Webhook/async patterns for long-running operations (Low)

Discoverability (DISC)

  1. DISC_001 OpenAPI 3.0+ used (High)
  2. DISC_002 Server URLs defined (Critical)
  3. DISC_003 Multiple environments documented (staging, prod) (Medium)
  4. DISC_004 API version in URL or header (Medium)
  5. DISC_005 CORS documented (Low)
  6. DISC_006 Health check endpoint exists (Low)

Workflow

Step 0: Pre-flight

  1. Find the spec: Look for OpenAPI files (**/openapi.{json,yaml,yml}, **/swagger.{json,yaml,yml}, **/*-api.{json,yaml,yml}). If none found, ask the user.
  2. Validate: Confirm parseable YAML/JSON with at least info and paths. If invalid, report errors and stop.
  3. Check MCP: Try getWorkspaces via Postman MCP.
    • MCP available: full analysis + Postman push capabilities
    • MCP unavailable: static spec analysis only. Note: "Postman MCP isn't configured. I can still analyze and fix your spec."

Step 1: Discover

Find specs locally and from Postman (if MCP available):

  • Local: **/openapi.{json,yaml,yml}, **/swagger.*, **/*-api.*
  • Postman: getAllSpecs + getSpecDefinition

If multiple specs found, list and ask which to analyze.

Step 2: Analyze

Read the spec and evaluate all 48 checks. For each:

  1. Examine relevant parts of the spec
  2. Count passing and failing items
  3. Assign pass/fail/partial status
  4. Calculate weighted score

Scoring formula:

  • Per check: weight * (passing_items / total_items) (skip N/A checks)
  • Per pillar: sum(weighted_scores) / sum(applicable_weights) * 100
  • Overall: sum(all_weighted_scores) / sum(all_applicable_weights) * 100

Severity weights: Critical = 4, High = 2, Medium = 1, Low = 0.5

Step 3: Present Results

Overall Score and Verdict:

Score: 67/100
Verdict: NOT AGENT-READY (need 70+ with no critical failures)

Pillar Breakdown:

Metadata:        ████████░░  82%
Errors:          ████░░░░░░  41%  <- Problem
Introspection:   ███████░░░  72%
Naming:          █████████░  91%
Predictability:  ██████░░░░  63%  <- Problem
Documentation:   ███░░░░░░░  35%  <- Problem
Performance:     █████░░░░░  52%
Discoverability: ████████░░  80%

Top 5 Priority Fixes (sorted by impact): For each, include:

  1. The check ID and what failed
  2. Why it matters for agents (concrete failure scenario)
  3. How to fix it (specific code example from their spec)

Step 4: Offer Next Steps

  1. "Want me to fix these?" - Walk through fixes one by one, editing the spec
  2. "Run again after fixes" - Re-analyze, show score improvement
  3. "Generate full report" - Save detailed markdown report to the project
  4. "Export to Postman" - Push improved spec, set up collection + environment + mock + docs

Fixing Issues

When the user says "fix these" or "improve my score":

  1. Start with highest-impact fix (highest severity x most endpoints affected)
  2. Read the relevant section of their spec
  3. Show the specific change with before/after
  4. Make the edit with user approval
  5. Move to next fix
  6. After all fixes, re-analyze to show new score

Postman MCP Integration

After analysis and fixes, if Postman MCP is available:

  1. Push spec: createSpec to store the improved spec
  2. Generate collection: generateCollection (async, poll for completion)
  3. Create environment: createEnvironment with base_url and auth variables
  4. Create mock: createMock for frontend development
  5. Run tests: runCollection to validate
  6. Publish docs: publishDocumentation to make docs public

From "broken API" to "fully operational Postman workspace" in one session.

Tone

  • Direct. "Your API scores 45%. That's not great. Here's what's dragging it down."
  • Specific. Always point to the exact check, endpoint, and fix.
  • Practical. Show the code change, not a REST theory lecture.
  • Encouraging when earned. "Your naming is solid at 91%. The errors pillar is what's killing you."

Quick Reference

User SaysWhat To Do
"Is my API agent-ready?"Discover specs, run analysis, present score
"Scan my project"Find all specs, summarize each
"What's wrong?"Show top 5 failures sorted by impact
"Fix it"Walk through fixes one by one, edit spec
"Run again"Re-analyze, show before/after comparison
"Generate report"Save detailed markdown report to project
"How do I get to 90%?"Calculate gap, show exactly which fixes get there
"Export to Postman"Push spec, generate collection, set up workspace

See references/pillars.md for the full pillar reference with detailed rationale.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

postman

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

Planning with files

Implements Manus-style file-based planning to organize and track progress on complex tasks. Creates task_plan.md, findings.md, and progress.md. Use when aske...

Registry SourceRecently Updated
228.4K
Profile unavailable
Coding

Nutrient Document Processing (Universal Agent Skill)

Universal (non-OpenClaw) Nutrient DWS document-processing skill for Agent Skills-compatible products. Best for Claude Code, Codex CLI, Gemini CLI, Cursor, Wi...

Registry SourceRecently Updated
0270
Profile unavailable