geo-technical

Technical SEO audit with GEO-specific checks — crawlability, indexability, security, performance, SSR, and AI crawler access

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "geo-technical" with this command: npx skills add zubair-trabzada/geo-seo-claude/zubair-trabzada-geo-seo-claude-geo-technical

GEO Technical SEO Audit

Purpose

Technical SEO forms the foundation of both traditional search visibility and AI search citation. A technically broken site cannot be crawled, indexed, or cited by any platform. This skill audits 8 categories of technical health with specific attention to GEO requirements — most critically, server-side rendering (AI crawlers do not execute JavaScript) and AI crawler access (many sites inadvertently block AI crawlers in robots.txt).

How to Use This Skill

  1. Collect the target URL (homepage + 2-3 key inner pages)
  2. Fetch each page using curl/WebFetch to get raw HTML and HTTP headers
  3. Run through each of the 8 audit categories below
  4. Score each category using the rubric
  5. Generate GEO-TECHNICAL-AUDIT.md with results

Category 1: Crawlability (15 points)

1.1 robots.txt Validity

  • Fetch https://[domain]/robots.txt
  • Check for syntactic validity: proper User-agent, Allow, Disallow directives
  • Check for common errors: missing User-agent, wildcards blocking important paths, Disallow: / blocking entire site
  • Verify XML sitemap is referenced: Sitemap: https://[domain]/sitemap.xml

1.2 AI Crawler Access (CRITICAL for GEO)

Check robots.txt for directives targeting these AI crawlers:

CrawlerUser-AgentPlatform
GPTBotGPTBotChatGPT / OpenAI
Google-ExtendedGoogle-ExtendedGemini / Google AI training
GooglebotGooglebotGoogle Search + AI Overviews
BingbotbingbotBing Copilot + ChatGPT (via Bing)
PerplexityBotPerplexityBotPerplexity AI
ClaudeBotClaudeBotAnthropic Claude
AmazonbotAmazonbotAlexa / Amazon AI
CCBotCCBotCommon Crawl (used by many AI models)
FacebookBotFacebookExternalHitMeta AI
BytespiderBytespiderTikTok / ByteDance AI
Applebot-ExtendedApplebot-ExtendedApple Intelligence

Scoring for AI crawler access:

  • All major AI crawlers allowed: 5 points
  • Some blocked but Googlebot + Bingbot allowed: 3 points
  • GPTBot or PerplexityBot blocked: 1 point (significant GEO impact)
  • Googlebot blocked: 0 points (fatal)

Important nuance: Blocking Google-Extended does NOT block Googlebot. Google-Extended only controls AI training data usage, not search indexing. However, blocking Google-Extended may reduce presence in AI Overviews. Recommend allowing Google-Extended unless there is a specific data licensing concern.

1.3 XML Sitemaps

  • Fetch sitemap (check robots.txt for location, or try /sitemap.xml, /sitemap_index.xml)
  • Validate XML syntax
  • Check for <lastmod> dates (should be present and accurate)
  • Count URLs — compare to expected number of indexable pages
  • Check for sitemap index if large site (50,000+ URLs per sitemap max)
  • Verify all sitemap URLs return 200 status codes (sample check)

1.4 Crawl Depth

  • Homepage = depth 0. Check that all important pages are reachable within 3 clicks (depth 3)
  • Pages at depth 4+ receive significantly less crawl budget and are less likely to be cited by AI
  • Check internal linking: are key content pages linked from the homepage or main navigation?

1.5 Noindex Management

  • Check for <meta name="robots" content="noindex"> on pages that SHOULD be indexed
  • Check for X-Robots-Tag: noindex HTTP headers
  • Common mistakes: noindex on paginated pages, category pages, or key landing pages

Category Scoring:

CheckPoints
robots.txt valid and complete3
AI crawlers allowed5
XML sitemap present and valid3
Crawl depth within 3 clicks2
No erroneous noindex directives2

Category 2: Indexability (12 points)

2.1 Canonical Tags

  • Every indexable page must have a <link rel="canonical" href="..."> tag
  • Canonical must point to itself (self-referencing) for the authoritative version
  • Check for conflicting canonicals (canonical in HTML vs. HTTP header)
  • Check for canonical chains (A canonicals to B, B canonicals to C — should be A to C)

2.2 Duplicate Content

  • Check for www vs. non-www (both should resolve, one should redirect)
  • Check for HTTP vs. HTTPS (HTTP should redirect to HTTPS)
  • Check for trailing slash consistency (pick one pattern and redirect the other)
  • Check for parameter-based duplicates (?sort=price creating duplicate pages)

2.3 Pagination

  • If paginated content exists, check for rel="next" / rel="prev" (note: Google ignores these as of 2019, but Bing still uses them)
  • Preferred: use rel="canonical" on paginated pages pointing to a view-all page or the first page
  • Ensure paginated pages are not noindexed if they contain unique content

2.4 Hreflang (international sites)

  • Check for <link rel="alternate" hreflang="xx"> tags
  • Validate: reciprocal hreflang (if page A points to page B, B must point back to A)
  • Validate: x-default fallback exists
  • Check for language/region code validity (ISO 639-1 / ISO 3166-1)

2.5 Index Bloat

  • Estimate number of indexed pages (check sitemap count, use site:domain.com estimate)
  • Compare indexed pages to actual valuable content pages
  • Flag if indexed pages significantly exceed content pages (index bloat from thin/duplicate/parameter pages)

Category Scoring:

CheckPoints
Canonical tags correct on all pages3
No duplicate content issues3
Pagination handled correctly2
Hreflang correct (if applicable)2
No index bloat2

Category 3: Security (10 points)

3.1 HTTPS Enforcement

  • Site must load over HTTPS
  • HTTP must redirect to HTTPS (301 redirect)
  • No mixed content warnings (HTTP resources on HTTPS pages)
  • SSL/TLS certificate must be valid and not expired

3.2 Security Headers

Check HTTP response headers for:

HeaderRequired ValuePurpose
Strict-Transport-Securitymax-age=31536000; includeSubDomainsForces HTTPS
Content-Security-PolicyAppropriate policyPrevents XSS
X-Content-Type-OptionsnosniffPrevents MIME sniffing
X-Frame-OptionsDENY or SAMEORIGINPrevents clickjacking
Referrer-Policystrict-origin-when-cross-origin or stricterControls referrer data
Permissions-PolicyAppropriate restrictionsControls browser features

Category Scoring:

CheckPoints
HTTPS enforced with valid cert4
HSTS header present2
X-Content-Type-Options1
X-Frame-Options1
Referrer-Policy1
Content-Security-Policy1

Category 4: URL Structure (8 points)

4.1 Clean URLs

  • URLs should be human-readable: /blog/seo-guide not /blog?id=12345
  • No session IDs in URLs
  • Lowercase only (no mixed case)
  • Hyphens for word separation (not underscores)
  • No special characters or encoded spaces

4.2 Logical Hierarchy

  • URL path should reflect site architecture: /category/subcategory/page
  • Flat where appropriate — avoid unnecessarily deep nesting
  • Consistent pattern across the site

4.3 Redirect Chains

  • Check for redirect chains (A redirects to B redirects to C)
  • Maximum 1 hop recommended (A redirects to C directly)
  • Check for redirect loops
  • All redirects should be 301 (permanent), not 302 (temporary), unless intentionally temporary

4.4 Parameter Handling

  • URL parameters should not create duplicate indexable pages
  • Use canonical tags or robots.txt Disallow for parameter variations
  • Configure parameter handling in Google Search Console and Bing Webmaster Tools

Category Scoring:

CheckPoints
Clean, readable URLs2
Logical hierarchy2
No redirect chains (max 1 hop)2
Parameter handling configured2

Category 5: Mobile Optimization (10 points)

Critical Context

As of July 2024, Google crawls ALL sites exclusively with mobile Googlebot. There is no desktop crawling. If your site does not work on mobile, it does not work for Google. Period.

5.1 Responsive Design

  • Check for <meta name="viewport" content="width=device-width, initial-scale=1">
  • Content must not require horizontal scrolling on mobile
  • No fixed-width layouts wider than viewport

5.2 Tap Targets

  • Interactive elements (buttons, links) must be at least 48x48 CSS pixels
  • Minimum 8px spacing between tap targets
  • Check that navigation is usable on mobile

5.3 Font Sizes

  • Base font size should be at least 16px
  • No text requiring zoom to read
  • Sufficient contrast ratio (WCAG AA: 4.5:1 for normal text, 3:1 for large text)

5.4 Mobile Content Parity

  • All content visible on desktop must also be visible on mobile
  • No hidden content behind "read more" toggles that Googlebot cannot expand (though Google has improved at expanding these as of 2025)
  • Images and media must load on mobile

Category Scoring:

CheckPoints
Viewport meta tag correct3
Responsive layout (no horizontal scroll)3
Tap targets appropriately sized2
Font sizes legible2

Category 6: Core Web Vitals (15 points)

2026 Metrics and Thresholds

Core Web Vitals use the 75th percentile of real user data (field data) as the benchmark. Lab data is useful for debugging but field data determines the ranking signal.

MetricGoodNeeds ImprovementPoorNotes
LCP (Largest Contentful Paint)< 2.5s2.5s - 4.0s> 4.0sMeasures loading — time until largest visible element renders
INP (Interaction to Next Paint)< 200ms200ms - 500ms> 500msReplaced FID in March 2024. Measures ALL interactions, not just first
CLS (Cumulative Layout Shift)< 0.10.1 - 0.25> 0.25Measures visual stability — unexpected layout movements

How to Assess Without CrUX Data

When real user data is unavailable, estimate from page characteristics:

  • LCP: Check largest above-fold element. Is it an image (check size/format)? Is it text (check web font loading)? Server response time (TTFB)?
  • INP: Check for heavy JavaScript on page. Long tasks (>50ms) block interactivity. Check for third-party scripts.
  • CLS: Check for images without explicit width/height. Check for dynamically inserted content above the fold. Check for web fonts causing layout shift (FOUT/FOIT).

Common LCP Fixes

  1. Optimize hero images: WebP/AVIF format, correct sizing, preload with <link rel="preload">
  2. Reduce server response time (TTFB < 800ms)
  3. Eliminate render-blocking CSS/JS
  4. Preconnect to critical third-party origins

Common INP Fixes

  1. Break up long tasks (>50ms) into smaller chunks using requestIdleCallback or scheduler.yield()
  2. Reduce third-party JavaScript
  3. Use content-visibility: auto for off-screen content
  4. Debounce/throttle event handlers

Common CLS Fixes

  1. Always include width and height attributes on images and videos
  2. Reserve space for ads and embeds with CSS aspect-ratio or explicit dimensions
  3. Use font-display: swap with size-adjusted fallback fonts
  4. Avoid inserting content above existing content after page load

Category Scoring:

CheckPoints
LCP < 2.5s5
INP < 200ms5
CLS < 0.15

Category 7: Server-Side Rendering (15 points) — CRITICAL FOR GEO

Why SSR Is Mandatory for AI Visibility

AI crawlers (GPTBot, PerplexityBot, ClaudeBot, etc.) do NOT execute JavaScript. They fetch the raw HTML and parse it. If your content is rendered client-side by React, Vue, Angular, or any other JavaScript framework, AI crawlers see an empty page.

Even Googlebot, which does execute JavaScript, deprioritizes JS-rendered content due to the additional crawl budget required. Google processes JS rendering in a separate "rendering queue" that can delay indexing by days or weeks.

Detection Method

  1. Fetch the page with curl (no JavaScript execution): curl -s [URL]
  2. Compare the raw HTML to the rendered DOM (via browser)
  3. If key content (headings, paragraphs, product info, article text) is MISSING from the curl output, the site relies on client-side rendering

What to Check

  • Main content text: Is the article body / product description / page content in the raw HTML?
  • Headings: Are H1, H2, H3 tags present in raw HTML?
  • Navigation: Is the main navigation server-rendered?
  • Structured data: Is JSON-LD in the raw HTML or injected by JavaScript?
  • Meta tags: Are title, description, canonical, OG tags in the raw HTML?
  • Internal links: Are navigation and content links in the raw HTML? (Critical for crawlability)

SSR Solutions to Recommend

FrameworkSSR Solution
ReactNext.js (SSR/SSG), Remix, Gatsby (SSG)
VueNuxt.js (SSR/SSG)
AngularAngular Universal
SvelteSvelteKit
GenericPrerender.io (prerendering service), Rendertron

Scoring Detail

  • All key content server-rendered: 15 points
  • Main content server-rendered but some elements JS-only: 10 points
  • Critical content requires JS (product info, article text): 5 points
  • Entire page is client-rendered (empty body in raw HTML): 0 points

Category Scoring:

CheckPoints
Main content in raw HTML8
Meta tags + structured data in raw HTML4
Internal links in raw HTML3

Category 8: Page Speed & Server Performance (15 points)

8.1 Time to First Byte (TTFB)

  • Target: < 800ms (ideally < 200ms)
  • Measure with curl: curl -o /dev/null -s -w 'TTFB: %{time_starttransfer}s\n' [URL]
  • If TTFB > 800ms: check server location, caching, database queries, CDN usage

8.2 Resource Optimization

  • Total page weight target: < 2MB (critical pages < 1MB)
  • Check for uncompressed resources (gzip/brotli compression should be enabled)
  • Check for unminified CSS and JavaScript
  • Check for unused CSS/JS (can represent 50%+ of downloaded bytes on many sites)

8.3 Image Optimization

  • Check image formats: WebP or AVIF preferred over JPEG/PNG
  • Check for oversized images (images larger than display size)
  • Check for lazy loading: images below fold should have loading="lazy"
  • Check for explicit dimensions (width/height attributes prevent CLS)
  • Above-fold images should NOT be lazy loaded (harms LCP)

8.4 Code Splitting and Lazy Loading

  • JavaScript should be code-split so each page only loads what it needs
  • Check for large JavaScript bundles (> 200KB compressed is a warning, > 500KB is critical)
  • Third-party scripts should load asynchronously (async or defer)
  • Check for render-blocking resources in <head>

8.5 Caching

  • Check Cache-Control headers on static resources (images, CSS, JS)
  • Static assets should have long cache times: max-age=31536000 (1 year) with content-hashed filenames
  • HTML pages should have shorter cache or no-cache with validation (ETag or Last-Modified)

8.6 CDN Usage

  • Check if static resources are served from a CDN (different domain or CDN-specific headers)
  • For global audience, CDN is critical for consistent performance
  • Check for CDN-specific headers: CF-Ray (Cloudflare), X-Cache (AWS CloudFront), X-Served-By (Fastly)

Category Scoring:

CheckPoints
TTFB < 800ms3
Page weight < 2MB2
Images optimized (format, size, lazy)3
JS bundles reasonable (< 200KB compressed)2
Compression enabled (gzip/brotli)2
Cache headers on static resources2
CDN in use1

IndexNow Protocol

What It Is

IndexNow is an open protocol that allows websites to notify search engines instantly when content is created, updated, or deleted. Supported by Bing, Yandex, Seznam, and Naver. Google does NOT support IndexNow but monitors the protocol.

Why It Matters for GEO

ChatGPT uses Bing's index. Bing Copilot uses Bing's index. Faster Bing indexing means faster AI visibility on two major platforms.

Implementation Check

  1. Check for IndexNow key file: https://[domain]/.well-known/indexnow-key.txt or similar
  2. Check if CMS has IndexNow plugin (WordPress: IndexNow plugin; many modern CMS platforms support it natively)
  3. If not implemented, recommend adding it with instructions

Overall Scoring

CategoryMax PointsWeight
Crawlability15Core foundation
Indexability12Core foundation
Security10Trust signal
URL Structure8Crawl efficiency
Mobile Optimization10Google requirement
Core Web Vitals15Ranking signal
Server-Side Rendering15GEO critical
Page Speed & Server15Performance
Total100

Score Interpretation

  • 90-100: Excellent — technically sound for both traditional SEO and GEO
  • 70-89: Good — minor issues to address but fundamentally solid
  • 50-69: Needs Work — significant technical debt impacting visibility
  • 30-49: Poor — major issues blocking crawling, indexing, or AI visibility
  • 0-29: Critical — fundamental technical failures requiring immediate attention

Output Format

Generate GEO-TECHNICAL-AUDIT.md with:

# GEO Technical SEO Audit — [Domain]
Date: [Date]

## Technical Score: XX/100

## Score Breakdown
| Category | Score | Status |
|---|---|---|
| Crawlability | XX/15 | Pass/Warn/Fail |
| Indexability | XX/12 | Pass/Warn/Fail |
| Security | XX/10 | Pass/Warn/Fail |
| URL Structure | XX/8 | Pass/Warn/Fail |
| Mobile Optimization | XX/10 | Pass/Warn/Fail |
| Core Web Vitals | XX/15 | Pass/Warn/Fail |
| Server-Side Rendering | XX/15 | Pass/Warn/Fail |
| Page Speed & Server | XX/15 | Pass/Warn/Fail |

Status: Pass = 80%+ of category points, Warn = 50-79%, Fail = <50%

## AI Crawler Access
| Crawler | User-Agent | Status | Recommendation |
|---|---|---|---|
| GPTBot | GPTBot | Allowed/Blocked | [Action] |
| Googlebot | Googlebot | Allowed/Blocked | [Action] |
[Continue for all AI crawlers]

## Critical Issues (fix immediately)
[List with specific page URLs and what is wrong]

## Warnings (fix this month)
[List with details]

## Recommendations (optimize this quarter)
[List with details]

## Detailed Findings
[Per-category breakdown with evidence]

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

Technical SEO Mastery

Conducts comprehensive technical SEO audits including crawlability, indexability, sitemap, robots.txt, Core Web Vitals, redirect chains, and international SE...

Registry SourceRecently Updated
4220Profile unavailable
General

geo

No summary provided by upstream source.

Repository SourceNeeds Review
Security

AetherCore v3.3

AetherCore v3.3.4 - Security-focused final release. High-performance JSON optimization with universal smart indexing for all file types. All security review...

Registry SourceRecently Updated
1471Profile unavailable
Security

QA & Testing Engine

Provides a comprehensive testing methodology for AI software, covering strategy design, unit, integration, and end-to-end tests with coverage and reporting g...

Registry SourceRecently Updated
8281Profile unavailable