Technical SEO audit methodology including crawlability, indexability, and Core Web Vitals analysis. Use when auditing pages or sites for technical SEO issues.
Installation
Details
Usage
After installing, this skill will be available to your AI coding assistant.
Verify installation:
skills listSkill Instructions
name: technical-audit description: Technical SEO audit methodology including crawlability, indexability, and Core Web Vitals analysis. Use when auditing pages or sites for technical SEO issues.
Technical Audit
When to Use
- Auditing pages for technical SEO issues
- Analyzing Core Web Vitals performance
- Checking schema markup implementation
- Validating crawlability and indexability
Audit Categories
1. Indexability
| Check | Requirement | Severity |
|---|---|---|
| Title Tag | Present, 50-60 chars, contains keyword | CRITICAL |
| Meta Description | Present, 150-160 chars | HIGH |
| Canonical Tag | Present, self-referencing or correct | HIGH |
| Robots Meta | No noindex on important pages | CRITICAL |
| Robots.txt | Not blocking important content | CRITICAL |
2. Content Structure
| Check | Requirement | Severity |
|---|---|---|
| H1 Tag | Exactly 1, contains keyword | CRITICAL |
| Heading Hierarchy | H1 -> H2 -> H3 (no skips) | HIGH |
| Word Count | Meets or exceeds competitor benchmark | MEDIUM |
| Content Uniqueness | No duplicate content issues | HIGH |
3. Core Web Vitals
| Metric | Good | Needs Improvement | Poor |
|---|---|---|---|
| LCP (Largest Contentful Paint) | <2.5s | 2.5s-4.0s | >4.0s |
| INP (Interaction to Next Paint) | <200ms | 200ms-500ms | >500ms |
| CLS (Cumulative Layout Shift) | <0.1 | 0.1-0.25 | >0.25 |
Measurement Methods:
- Chrome DevTools MCP (preferred)
- PageSpeed Insights API
- Lighthouse CLI
- Manual measurement via web.dev
4. Schema Markup
| Page Type | Recommended Schema |
|---|---|
| Article/Blog | Article, BlogPosting |
| FAQ page | FAQPage |
| How-to guide | HowTo |
| Product page | Product |
| Local business | LocalBusiness |
| Person/Author | Person |
5. Links
| Check | Requirement | Severity |
|---|---|---|
| Internal Links | Minimum 3 per page | HIGH |
| Broken Links | 0 | CRITICAL |
| External Links | At least 1 authoritative | LOW |
| Orphan Pages | 0 (all pages linked from somewhere) | MEDIUM |
Audit Process
Step 1: Fetch Page
# Use WebFetch or curl
curl -s "$URL" > page.html
Step 2: Parse Structure
- Extract title, meta description, canonical
- Map heading hierarchy
- Count words
- List all links
Step 3: Analyze Performance
- Use PageSpeed Insights API or Chrome DevTools MCP
- Document all Core Web Vitals
- Note specific issues (large images, render-blocking JS)
Step 4: Check Schema
- Look for JSON-LD in page source
- Validate using Google Rich Results Test
- Note missing or incomplete properties
Step 5: Score and Report
- Calculate overall score (0-100)
- List all issues by severity
- Provide specific fix recommendations
Output Format
## Technical SEO Audit Report
**URL**: {url}
**Date**: {date}
**Overall Score**: {score}/100
### Core Web Vitals
| Metric | Value | Status |
|--------|-------|--------|
| LCP | {value}s | GOOD/NEEDS IMPROVEMENT/POOR |
| INP | {value}ms | GOOD/NEEDS IMPROVEMENT/POOR |
| CLS | {value} | GOOD/NEEDS IMPROVEMENT/POOR |
**Measurement Method**: {Chrome DevTools MCP | PageSpeed API | Lighthouse | Manual}
### Issues Found
**CRITICAL ({count})**:
1. {issue} - {location} - {fix recommendation}
**HIGH ({count})**:
1. {issue} - {location} - {fix recommendation}
**MEDIUM ({count})**:
1. {issue} - {location} - {fix recommendation}
### Recommendations
1. {priority fix 1}
2. {priority fix 2}
3. {priority fix 3}
More by MadAppGang
View allChoose optimal external AI models for code analysis, bug investigation, and architectural decisions. Use when consulting multiple LLMs via claudish, comparing model perspectives, or investigating complex Go/LSP/transpiler issues. Provides empirically validated model rankings (91/100 for MiniMax M2, 83/100 for Grok Code Fast) and proven consultation strategies based on real-world testing.
CRITICAL - Guide for using Claudish CLI ONLY through sub-agents to run Claude Code with OpenRouter models (Grok, GPT-5, Gemini, MiniMax). NEVER run Claudish directly in main context unless user explicitly requests it. Use when user mentions external AI models, Claudish, OpenRouter, or alternative models. Includes mandatory sub-agent delegation patterns, agent selection guide, file-based instructions, and strict rules to prevent context window pollution.
MANDATORY tracking protocol for multi-model validation. Creates structured tracking tables BEFORE launching models, tracks progress during execution, and ensures complete results presentation. Use when running 2+ external AI models in parallel. Trigger keywords - "multi-model", "parallel review", "external models", "consensus", "model tracking".
XML tag structure patterns for Claude Code agents and commands. Use when designing or implementing agents to ensure proper XML structure following Anthropic best practices.