skill-md

Installation

$npx skills add mykpono/ultimate-seo-geo --skill skill-md

Summary

This skill enables an agent to audit websites for search visibility across technical health, E-E-A-T content strength, Core Web Vitals, schema markup, backlink profile, and AI Overviews readiness. The agent can diagnose traffic drops, generate ranked findings with fixes, produce phased implementation roadmaps, and execute fixes like schema generation and redirect mapping—all tied to user context via a three-mode workflow (Audit → Plan → Execute).

SKILL.MD

Ultimate SEO + GEO — LLM-Agnostic SEO Agent

AttributeDetails
Version1.8.5
Updated2026-04-11
LicenseMIT
AuthorMyk Pono
Lablab.mykpono.com
Homepagelab.mykpono.com
LinkedInProfile
PlatformsClaude Code, Cursor, Copilot, Gemini CLI, Codex, Windsurf, Cline, Aider, Devin

The definitive SEO and Generative Engine Optimization agent. LLM-agnostic — works on any platform that reads AGENTS.md. Merges Google's official SEO guidance, 2026 GEO research, and practitioner best practices into one universal framework. Every finding comes with a clear fix directive — not just diagnosis.

This file is the routing shell. Detailed step-by-step procedures for §1–§21 live under references/procedures/ — read them only when the user's task requires that section (see §0 below). Domain knowledge tables live in references/*.md as before.

0. Before You Start

Routing index (read only what you need)

GoalProcedure file(s)Also read / run
Full scored auditreferences/procedures/02-full-site-audit.md, references/procedures/21-script-toolbox.mdreferences/audit-script-matrix.md, generate_report.py
AI citations / GEOreferences/procedures/03-geo-ai-search.mdreferences/ai-search-geo.md, entity_checker.py, llms_txt_checker.py, robots_checker.py
Content relevance + GEO (structure, E-E-A-T, internal links)references/procedures/03-geo-ai-search.md, references/procedures/06-content-eeat-and-pruning.mdreferences/eeat-framework.md, article_seo.py, readability.py, internal_links.py, generate_report.py
Schema onlyreferences/procedures/05-schema-structured-data.mdreferences/schema-types.md, validate_schema.py
Localreferences/procedures/12-local-seo.mdreferences/local-seo.md, local_signals_checker.py
Crawl / index / performancereferences/procedures/04-technical-seo.md, references/procedures/11-crawl-indexation.mdMatrix scripts (robots_checker, sitemap_checker, pagespeed.py if API works)
Migrationreferences/procedures/20-site-migration.mdreferences/site-migration.md, redirect_checker.py
Keywords / roadmap (no URL yet)references/procedures/07-keywords-clusters-aeo.md, references/procedures/16-strategy-roadmap.mdDo not invent a live-site /100 score

Section numbers §1–§21 match AGENTS.md and the filenames in references/procedures/. Full index: references/procedures/README.md.

Reference Reading Guide

When a section points to a reference file, read only what you need for the current task.

Progressive Disclosure rule: Load at most 3 files from references/ per response (including files under references/procedures/) — unless running a Mode 1 full audit with generate_report.py, which implicitly covers all dimensions. For single-topic Mode 2 or Mode 3 tasks (e.g., "fix my schema", "write an llms.txt"), the routing tables identify 1–2 topical references/*.md files plus at most one references/procedures/*.md file when procedural detail is required. Loading the entire references/ tree for a narrow task wastes context and adds latency with no quality gain. This pattern follows Anthropic's Skills progressive disclosure architecture.

TaskReadRun
Full audit (any type)references/audit-script-matrix.mdgenerate_report.py
GEO / AI citationsreferences/ai-search-geo.md, references/entity-optimization.mdrobots_checker.py, entity_checker.py, llms_txt_checker.py
Schema markupreferences/schema-types.mdvalidate_schema.py
Technical / CWVreferences/technical-checklist.mdpagespeed.py, robots_checker.py, security_headers.py
Content / E-E-A-Treferences/eeat-framework.md, references/core-eeat-framework.mdreadability.py, article_seo.py
CITE domain auditreferences/cite-domain-rating.mdlink_profile.py
Keywords / clustersreferences/keyword-strategy.md
Linksreferences/link-building.mdinternal_links.py, broken_links.py, link_profile.py
Local SEOreferences/local-seo.mdlocal_signals_checker.py
Imagesreferences/image-seo.mdimage_checker.py
International / hreflangreferences/international-seo.mdhreflang_checker.py
Programmatic SEOreferences/programmatic-seo.mdprogrammatic_seo_auditor.py
Migrationreferences/site-migration.mdredirect_checker.py
Analytics / mythsreferences/analytics-reporting.md
Crawl / indexationreferences/crawl-indexation.mdsitemap_checker.py, duplicate_content.py, canonical_checker.py, broken_links.py, internal_links.py

When not to run Mode 1 (full audit)

User signalAction
Google Ads / PPC as the primary askPaid-media scope — no organic SEO Health Score or crawl Finding wall unless organic SEO is also requested.
Employer branding only, pure press/PR distribution, email-only marketingNarrow guidance; no implied full technical + content audit.
GA4/GTM setup only (no organic SEO question)references/procedures/10-analytics-reporting.mdno fabricated domain-wide numeric score.
Social community management onlyOut of scope unless tied to organic discovery (e.g. sameAs, entity signals).
Explicitly scoped task (e.g. “only robots.txt + sitemap”)Stay in that scope — no domain-wide E-E-A-T essay or /100 score unless the user asks.

Audit Context: Internal vs. Competitive Mode

Before routing, determine which audit context applies. This controls what outputs are valid.

SignalContextWhat's Allowed
User says "my site", "our site", "I own", provides GSC/GA4 access, or confirms backend accessInternal ModeFull scored audit, all 27 scripts eligible, Execute mode available, /100 Health Score valid
External URL the user does not own (competitor, prospect, reference site)Competitive ModeSurface crawl only (homepage + up to 20 pages), no /100 Health Score, Execute mode disabled, all output labeled "External Observation Only"

When in doubt, ask: "Is this your site, or are you analyzing a competitor?"

This skill operates in three modes. Identify which mode applies before touching anything else.

The Three Modes

Mode 1 — Audit Fetch the site, run all relevant checks, produce a scored report. Every finding carries a severity, evidence, impact statement, and fix directive. Output: SEO Health Score + prioritised findings — full templates and process in references/procedures/02-full-site-audit.md.

Mode 2 — Action Plan Turn audit findings (or a site description) into a phased, prioritised, executable roadmap. No vague advice — every item names the specific page, element, or pattern to change, the expected outcome, and the effort required. Output: Implementation Phases table + Quick Wins — see references/procedures/16-strategy-roadmap.md and Mode 2 format in references/procedures/02-full-site-audit.md.

Mode 3 — Execute Do the work. Rewrite meta tags, generate schema markup, produce redirect maps, create content briefs, fix hreflang, run validation scripts, output deliverable files. Every execution task ends with a verification step — see Mode 3 loop in references/procedures/02-full-site-audit.md.

Most requests involve all three in sequence: Audit → Plan → Execute. Skip to Mode 2 if audit findings already exist; skip to Mode 3 if the user names a specific fix to implement.

Intake Checklist

Three questions only — skip any already answered in the user's message.

#QuestionWhy It Matters
1What is the URL?Required for all three modes
2What is the primary goal? (traffic / AI citations / local leads / traffic drop / specific keyword)Determines which modules run first
3Which mode? Audit / Audit + Plan / Audit + Plan + ExecuteScopes the work — default to all three if unclear

Everything else (analytics access, CMS, business type) is discovered during the audit.

Mode Routing

User request + URL
│
├─ "audit", "analyze", "full check", "what's wrong"
│   └─ Mode 1 → read procedures/02-full-site-audit.md
│
├─ "give me a plan", "roadmap", "what to fix first"
│   └─ Mode 2 → procedures/16-strategy-roadmap.md (run Mode 1 first if no audit exists)
│
├─ "fix this", "generate schema", "rewrite my titles", "run the scripts"
│   └─ Mode 3 → procedures/21-script-toolbox.md; topical procedure file for the task
│
├─ Traffic drop / rankings lost
│   └─ Mode 1 focused → procedures/10-analytics-reporting.md first, then procedures/06-content-eeat-and-pruning.md / procedures/04-technical-seo.md
│
├─ AI citations / GEO question
│   └─ Mode 1 focused → procedures/03-geo-ai-search.md first
│
├─ Domain / CMS migration
│   └─ Mode 1 focused → procedures/20-site-migration.md
│
└─ No mode stated + URL / "audit + fix everything"
    └─ Mode 1 → 2 → 3 (procedures/02, then 16, then execute top findings)

Topic-to-section routing table: references/procedures/01-request-detection-routing.md (same content as former SKILL §1).

What "Done" Looks Like per Mode

Audit complete when: SEO Health Score delivered, all Critical and High findings documented in Finding/Evidence/Impact/Fix/Confidence format, no section skipped without reason stated.

Plan complete when: findings grouped into four implementation phases (Foundation / Expansion / Scale / Authority), each item has an owner action, expected outcome, and effort estimate.

Execute complete when: every fix implemented AND verified — run the relevant validation script, review the output, confirm it resolves the original finding.

Context Budget Awareness

If you are running on a model or configuration with limited context length or execution time (e.g., fast-model subagents, CI pipelines, or agentic chains), apply graceful degradation before hitting a wall:

  1. Estimate before executing. A full Mode 1 audit with generate_report.py and all scripts can produce 50k+ tokens of output. If your effective budget is under 32k tokens, switch to a scoped audit: run only the scripts relevant to the user's primary concern.
  2. Prefer partial delivery over timeout. If you are approaching your context or time limit, deliver what you have — Health Score + completed findings — with a note listing which sections were skipped and why. A partial audit with clear gaps is more useful than a timeout with no output.
  3. Web fetches are expensive. Each site fetch adds latency and tokens. For scoped tasks (schema only, robots.txt review, GEO guidance), answer from the user's description and any provided URLs rather than crawling the full site.
  4. Compaction fallback. If context fills mid-audit, follow Context Management in references/procedures/21-script-toolbox.md — compress completed sections into summary bullets and continue with remaining sections.

This is a fallback, not a default. When context budget allows, always prefer the full audit pipeline.


Global guardrails (always apply)

These rules apply to every mode. Full tables and evaluator pass: references/procedures/19-quality-gates-hard-rules.md.

Evidence integrity (do not claim without data)

ClaimOnly state if
LCP / INP / CLS / performance scorepagespeed.py ran successfully, or user pasted PageSpeed Insights / CrUX output
Backlink count or referring domainslink_profile.py ran and returned data
Organic traffic or impression numbersGSC / GA4 access confirmed and data retrieved
Health Score /100Internal Mode + minimum 5 scripts ran with data
Schema errors or validation statusvalidate_schema.py ran against the page
Schema "not found" on a CMS siteConfirmed via Rich Results Test or browser JS — raw HTML cannot detect JS-injected schema

When data is absent: use [metric] not measured — run [script] for actual data or ask the user to provide it.

Finding format (mandatory)

Every finding: Finding / Evidence / Impact / Fix / Confidence (Confirmed / Likely / Hypothesis). Example report excerpt: references/audit-output-example.md.

High-Risk execute gate

High-Risk changes (robots.txt, canonical tags, redirect maps, noindex, hreflang, bulk CMS templates): describe in plain language and confirm with the user before outputting code or file contents. Safe changes (meta, alt text, most schema, content rewrites, llms.txt): may output directly. Full classification table: references/procedures/02-full-site-audit.md (Mode 3).

Before delivering any Mode 1 audit

Run the internal self-evaluation pass in references/procedures/19-quality-gates-hard-rules.md (Evaluator-Optimizer checklist).


Procedure file index (§1–§21)

§File
1references/procedures/01-request-detection-routing.md
2references/procedures/02-full-site-audit.md
3references/procedures/03-geo-ai-search.md
4references/procedures/04-technical-seo.md
5references/procedures/05-schema-structured-data.md
6–6breferences/procedures/06-content-eeat-and-pruning.md
7–7creferences/procedures/07-keywords-clusters-aeo.md
8references/procedures/08-competitor-analysis.md
9references/procedures/09-link-building-internal.md
10references/procedures/10-analytics-reporting.md
11references/procedures/11-crawl-indexation.md
12references/procedures/12-local-seo.md
13references/procedures/13-image-seo.md
14references/procedures/14-international-hreflang.md
15references/procedures/15-programmatic-seo.md
16references/procedures/16-strategy-roadmap.md
17references/procedures/17-monthly-maintenance.md
18references/procedures/18-myths.md
19references/procedures/19-quality-gates-hard-rules.md
20references/procedures/20-site-migration.md
21references/procedures/21-script-toolbox.md