blog-geo

AI citation optimization audit scoring blog posts for ChatGPT, Perplexity, and Google AI Overview citability. Evaluates passage-level citability, Q&A formatting, entity clarity, structured data, and AI crawler accessibility. Generates citation capsules and a 0-100 AI Citation Readiness score. Use when user says "geo", "ai citation", "ai optimization", "citation audit", "aeo", "perplexity optimization", "chatgpt citation".

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "blog-geo" with this command: npx skills add megastep/codex-skills/megastep-codex-skills-blog-geo

Blog GEO -- AI Citation Optimization Audit

Scores blog posts for AI citation readiness across ChatGPT, Perplexity, and Google AI Overviews. Generates citation capsules and a 0-100 AI Citation Readiness score with platform-specific recommendations.

Key Research Data

Reference these benchmarks throughout the audit:

  • Only 12% of sources cited match across ChatGPT, Perplexity, and AI Overviews
  • 80% of LLM citations don't rank in Google's top 100
  • Brands 6.5x more likely cited through third-party sources
  • 120-180 word sections get 70% more ChatGPT citations
  • Comparison tables with <thead> achieve 47% higher AI citation rates
  • Content freshness: 76.4% of top citations updated within 30 days

Audit Process

Step 1: Read Content

Extract from the blog post:

  • Full content text and word count
  • Heading structure (H1, H2, H3 hierarchy)
  • Individual paragraphs and their word counts
  • FAQ sections (if present)
  • Schema markup (JSON-LD, microdata, RDFa)
  • robots.txt mentions or meta robots directives
  • Any TL;DR or summary boxes
  • Comparison tables and their HTML structure
  • Numbered/ordered lists
  • Definition-style formatting

Step 2: Passage-Level Citability (4 pts)

Check each section between headings for AI-extractable passages:

CheckCriteria
Word countEach section contains 120-180 word self-contained passages
Context independenceEach passage makes sense extracted from surrounding context
Claim structurePassages contain: specific claim + supporting evidence + source attribution
CompletenessPassage answers a question without requiring reader to read adjacent sections

Scoring: Count passages meeting all criteria vs total sections.

  • 4 pts: 80%+ sections have citable passages
  • 3 pts: 60-79%
  • 2 pts: 40-59%
  • 1 pt: 20-39%
  • 0 pts: <20%

Step 3: Q&A Formatting (3 pts)

Check heading format and answer structure:

CheckCriteria
Question headings60-70% of H2s are phrased as questions
Answer-first formatOpening paragraph under each H2 provides a direct answer
FAQ sectionDedicated FAQ section with structured question-answer pairs

Scoring:

  • 3 pts: All three criteria met
  • 2 pts: Two criteria met
  • 1 pt: One criterion met
  • 0 pts: None met

Step 4: Entity Clarity (3 pts)

Check topic consistency and disambiguation:

CheckCriteria
Canonical topicOne unambiguous primary topic per page
Consistent namingSame entity name used throughout (no confusing synonyms)
Intro statementClear topic statement in the introduction paragraph
Title-content matchTitle accurately reflects the content focus

Scoring:

  • 3 pts: All four criteria met
  • 2 pts: Three criteria met
  • 1 pt: One or two criteria met
  • 0 pts: None met

Step 5: Content Structure for Extraction (3 pts)

Check for AI-extractable content patterns:

CheckCriteria
TL;DR box40-60 word standalone summary present at top
Comparison tablesTables with proper HTML <thead> (47% higher citation rate)
Ordered listsNumbered lists for processes and step-by-step instructions
Definition formattingKey terms formatted with clear definition patterns
Citation capsules40-60 word definitive statements in each major section

Scoring:

  • 3 pts: 4-5 elements present
  • 2 pts: 3 elements present
  • 1 pt: 1-2 elements present
  • 0 pts: None present

Step 6: AI Crawler Accessibility (2 pts)

Check technical requirements for AI crawler indexing:

CheckCriteria
Static HTMLContent rendered in static HTML, not behind JavaScript
robots.txtAllows AI crawlers: GPTBot, ChatGPT-User, ClaudeBot, PerplexityBot
Schema in HTMLSchema markup in static HTML, not JS-injected
Page sizeReasonable page size within AI crawler limits

Scoring:

  • 2 pts: All criteria met
  • 1 pt: Most criteria met but one issue
  • 0 pts: Multiple issues blocking AI crawlers

Step 7: Platform-Specific Analysis

Evaluate the post for each AI platform's citation preferences:

ChatGPT

  • Favors "Best X" listicles (43.8% of citations)
  • Prefers well-cited, authoritative content
  • Recency matters -- recent updates get priority
  • Domain authority influences citation likelihood

Perplexity

  • Favors Reddit sources (6.6% of all citations)
  • Rapid content decay: 2-3 day citation window
  • Freshness is the most critical factor
  • Community-validated content preferred

Google AI Overviews

  • Favors Google properties (23% of citations)
  • High Domain Rating strongly correlated with citation
  • Present in 49% of SERPs
  • Prefers content that already ranks well organically

For each platform, provide:

  • Current citability rating (High / Medium / Low)
  • Specific improvements to increase citation likelihood
  • Content format recommendations

Step 8: Generate Citation Capsules

For each H2 section in the post, write a citation capsule:

  • Length: 40-60 words, self-contained
  • Structure: Specific claim + data point + source attribution
  • Purpose: A passage AI could directly quote as a citation
  • Format: Present as a suggested addition the author can embed

Example:

According to [Source], [specific claim with number]. This represents
[context/comparison], making it [significance]. [Supporting detail
that reinforces the claim].

Generate one capsule per H2 section. Label each with the section heading it belongs under.

Step 9: Calculate AI Citation Readiness Score (0-100)

Map the 15-point subcategory scores to a 0-100 display score:

CategoryRaw PointsDisplay WeightMax Display Score
Passage-Level Citability/4x6.7527
Q&A Formatting/3x6.6720
Entity Clarity/3x6.6720
Content Structure/3x6.6720
AI Crawler Accessibility/2x6.513
Total/15100

Rating thresholds:

  • 90-100: Excellent -- highly citable by AI systems
  • 70-89: Good -- citable with minor improvements
  • 50-69: Needs Work -- significant gaps in citability
  • Below 50: Poor -- major restructuring needed

Step 10: Generate Report

Output the following report:

## AI Citation Readiness Report: [Title]

**AI Citation Readiness Score: [X]/100** -- [Rating]

### Score Breakdown
| Category | Raw | Display | Max |
|----------|-----|---------|-----|
| Passage-Level Citability | X/4 | X | 27 |
| Q&A Formatting | X/3 | X | 20 |
| Entity Clarity | X/3 | X | 20 |
| Content Structure | X/3 | X | 20 |
| AI Crawler Accessibility | X/2 | X | 13 |
| **Total** | **X/15** | **X** | **100** |

### Per-Section Citability Analysis
| Section (H2) | Word Count | Self-Contained | Claim+Evidence | Citable |
|---------------|-----------|----------------|----------------|---------|
| [heading] | [N] | Yes/No | Yes/No | Yes/No |

### Platform-Specific Optimization
#### ChatGPT
- [specific recommendations]

#### Perplexity
- [specific recommendations]

#### Google AI Overviews
- [specific recommendations]

### Generated Citation Capsules

#### [H2 Section 1]
> [40-60 word citation capsule]

#### [H2 Section 2]
> [40-60 word citation capsule]

### Technical Recommendations
- [ ] [Technical fix with specifics]

### Priority Action Items
1. [Most impactful improvement]
2. [Second most impactful]
3. [Third most impactful]

Run `$blog analyze <file>` for full content quality scoring.

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Security

ads-audit

No summary provided by upstream source.

Repository SourceNeeds Review
Security

blog-audit

No summary provided by upstream source.

Repository SourceNeeds Review
Security

seo-audit

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

ads-competitor

No summary provided by upstream source.

Repository SourceNeeds Review