Two weeks ago we published a 60-tag, 14-category audit covering every meta tag modern sites need in 2026. The category that got the most reader questions was Answer Engine Optimization — the new layer of signals that decides whether AI assistants cite your page or skip it. The original piece argued that AEO matters. This one shows you how.

This is the tactical follow-up. Working JSON-LD you can paste into your <head>, before/after conversions from ordinary prose to citation-ready markup, and the specific patterns the four biggest answer engines — ChatGPT, Claude, Perplexity, and Google's AI Overviews — preferentially pull from.

If you implemented every Core SEO and Open Graph recommendation from the original audit but skipped the AEO layer, your pages are indexed. They just aren't cited. In 2026, those are two very different things.

Why AI Assistants Cite Some Pages and Skip Others

Traditional search rewards pages that rank. Answer engines reward pages that are easy to assemble an answer from. That's a different optimization target, and it changes everything about how you structure a page.

When Perplexity pulls sources for a query, or when Google's AI Overviews assembles a multi-sentence answer with citations underneath, the system isn't reading your page the way a user does. It's looking for extractable units: a clearly scoped question, a concise answer directly adjacent to it, and strong signals that the answer is authoritative. Schema markup is a structural hint that says "here is a question, here is its answer, and the relationship between them is explicit."

Prose paragraphs require the model to do extraction work. Schema gives it extraction for free. When two pages have comparable answers, the one that's already in extractable form is cheaper to cite — and that's the one that wins the citation.

The Citation Economics

Think of an AI assistant as a lazy editor with a deadline. It will cite the page that required the least work to turn into a sourced sentence. Schema isn't a ranking bonus — it's a reduction in the cost of citing you. All else equal, cheaper citations win.

FAQ Schema: The Format AI Overviews Love

FAQ schema is the single highest-leverage AEO signal. It's directly citable, universally supported across answer engines, and easy to add to almost any content page. Here's the minimal working shape:

<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [
    {
      "@type": "Question",
      "name": "What is Answer Engine Optimization?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "Answer Engine Optimization is the practice of structuring content so AI assistants and answer engines cite it when answering user questions. It relies on schema markup, concise answer blocks, and strong entity signals."
      }
    },
    {
      "@type": "Question",
      "name": "How is AEO different from SEO?",
      "acceptedAnswer": {
        "@type": "Answer",
        "text": "SEO optimizes for ranking in blue-link search results. AEO optimizes for being cited inside AI-generated answers. The two overlap but are not the same — a page can rank #1 and still be skipped by AI Overviews if its content isn't extractable."
      }
    }
  ]
}
</script>

A few non-obvious rules that the Schema.org spec doesn't enforce but the answer engines do:

Before / After: Prose to FAQ Schema

Most pages already contain good answers — they're just buried in paragraphs. The job isn't to write new content; it's to restructure what's already there. Here's a typical before:

<p>Many developers wonder about canonical tags. The short
version is that a canonical tag tells search engines which
URL is the authoritative version of a page when multiple
URLs serve similar content. Without one, your staging site
might compete with production.</p>

That paragraph has a good answer inside it, but no AI assistant is going to easily extract "what does a canonical tag do" from a sentence that starts with "Many developers wonder about canonical tags." Here's the same information restructured for AEO:

<h3>What does a canonical tag do?</h3>
<div class="answer-block">
  <p>A canonical tag tells search engines which URL is the
  authoritative version of a page when multiple URLs serve
  similar content. Without one, staging environments, tracking
  parameters, or duplicate paths can compete with your canonical
  page in search results.</p>
</div>

<!-- And in the <head>: -->
<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "FAQPage",
  "mainEntity": [{
    "@type": "Question",
    "name": "What does a canonical tag do?",
    "acceptedAnswer": {
      "@type": "Answer",
      "text": "A canonical tag tells search engines which URL is the authoritative version of a page when multiple URLs serve similar content. Without one, staging environments, tracking parameters, or duplicate paths can compete with your canonical page in search results."
    }
  }]
}
</script>

Three things changed. The heading became the exact conversational question a user would type. The answer moved directly under the heading as its own block. The JSON-LD mirrors the visible content word-for-word. That's the whole pattern, repeated across every major question your page answers.

The 40-60 Word Answer Block Pattern

The heading-plus-answer-block structure above is the single highest-leverage formatting pattern in AEO. It's what Google pulls featured snippets from, what AI Overviews lifts verbatim, and what Perplexity wraps in a citation. Here's the template:

▸ Anatomy of an AEO Answer Block

One question-format heading. One 40-60 word answer directly underneath, written as a complete thought that could stand alone if quoted. Then — and only then — expand with context, examples, and caveats in regular paragraphs below. The answer block is the citation bait. Everything after it is the rest of the article.

The reason for the 40-60 word target is mechanical: it's the size window most AI assistants use when lifting a single-paragraph citation. Shorter answers don't contain enough context to stand alone. Longer answers get truncated mid-thought. Land inside the window and the whole block travels together.

HowTo Schema: Structured Instructions for Step-by-Step Queries

If FAQ schema owns the "what is" and "why does" queries, HowTo schema owns the "how do I" queries. Any tutorial, step-by-step guide, recipe, workflow, or numbered procedure belongs in HowTo markup. The minimal shape:

<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "HowTo",
  "name": "How to audit your meta tags with curl",
  "description": "A three-step process for extracting and validating meta tags on any page using only standard Unix tools.",
  "totalTime": "PT10M",
  "step": [
    {
      "@type": "HowToStep",
      "position": 1,
      "name": "Extract the head section",
      "text": "Run curl with sed to isolate the <head> block from the raw HTML.",
      "url": "https://example.com/audit-guide#step-1"
    },
    {
      "@type": "HowToStep",
      "position": 2,
      "name": "Validate structured data",
      "text": "Pipe any embedded JSON-LD blocks into python3 -m json.tool to verify they parse.",
      "url": "https://example.com/audit-guide#step-2"
    },
    {
      "@type": "HowToStep",
      "position": 3,
      "name": "Check image dimensions",
      "text": "Download the og:image URL and verify it meets the 1200x630 minimum for Facebook and LinkedIn.",
      "url": "https://example.com/audit-guide#step-3"
    }
  ]
}
</script>

HowTo schema carries some specific requirements that trip people up:

Speakable Schema: The Voice Answer Engine Opportunity

Speakable is the least-implemented and most-underrated AEO signal in 2026. It tells voice assistants — Google Assistant, Alexa, Siri, and the new generation of AI voice agents — which specific parts of your page they should read aloud when your content is the source of a voice answer. Most sites have zero Speakable markup, which means voice assistants either skip them entirely or read something random from the page.

<script type="application/ld+json">
{
  "@context": "https://schema.org",
  "@type": "WebPage",
  "name": "What does a canonical tag do?",
  "speakable": {
    "@type": "SpeakableSpecification",
    "cssSelector": [
      ".article-subtitle",
      ".answer-block"
    ]
  }
}
</script>

The cssSelector array points to elements on your page that should be read aloud. The pattern that works best: mark the subtitle (a one-sentence summary) and the answer blocks — short, self-contained text that sounds natural when spoken. Do not mark entire article bodies. Voice assistants will either truncate aggressively or pick the most quotable fragment, and you want to control which fragment that is.

The xpath variant is also valid and sometimes more precise:

"speakable": {
  "@type": "SpeakableSpecification",
  "xpath": [
    "/html/head/title",
    "/html/body//div[@class='answer-block'][1]"
  ]
}

FAQ vs. Q&A Schema: Don't Confuse Them

These are two different schema types with different use cases, and mixing them up is one of the most penalized AEO errors. Here's the distinction:

SCHEMA WHO WRITES THE ANSWERS CORRECT USE CASE
FAQPage You (site owner, authoritative) Product FAQs, documentation, knowledge base articles, any page where you provide both the question and the definitive answer.
QAPage Users (community-contributed) Forum threads, Stack Overflow-style Q&A, community support pages where one question has multiple user-submitted answers.

If you put QAPage schema on a product FAQ, Google will either ignore it or flag it as spam. If you put FAQPage schema on a forum thread, answer engines won't trust any of the answers because they know FAQPage implies editorial authority. Use the right one.

Testing Your Schema Before You Ship

Schema errors usually surface silently. Your markup parses, your page renders, but the answer engines quietly ignore your content for weeks before you notice the traffic gap. Build validation into your workflow:

Validate JSON Syntax Locally

# Extract all JSON-LD blocks and pipe through json.tool
curl -s https://yoursite.com/page | \
  grep -oP '<script type="application/ld\+json">\K.*?(?=</script>)' | \
  while read -r block; do
    echo "$block" | python3 -m json.tool > /dev/null && echo "OK" || echo "INVALID"
  done

Validate Against Schema.org

Google's Rich Results Test will tell you if your schema qualifies for rich results. Schema.org's validator gives you the raw schema validation without the Google-specific filters. Run both — they catch different classes of errors.

Test With the Answer Engines Directly

The fastest real-world test: paste your URL into Perplexity, ChatGPT, and Claude and ask a question your page should answer. If all three cite you, your AEO is working. If only one does, you have partial coverage. If none do, your schema is invisible, your content is hard to extract, or both.

The Invisible-Markup Trap

The most common failure mode: schema that claims questions and answers that don't appear anywhere on the visible page. Developers generate FAQ schema from a content management system, paste it into the <head>, and never check that the visible HTML matches. Answer engines notice within days and quietly stop trusting your markup. Always render the visible FAQ section from the same source as the JSON-LD.

What AI Assistants Actually Cite — Observed Patterns

After running thousands of test queries across the four major answer engines, some consistent patterns emerge. These aren't documented preferences — they're empirical behavior:

The Seven AEO Failures That Kill Citations

These are the recurring mistakes we see across real-world audits, ranked by how often they show up:

1. Invisible schema. Questions and answers exist in JSON-LD but nowhere on the rendered page. The single most common failure and the single biggest penalty.

2. Over-schema. Trying to claim the same content is simultaneously an Article, FAQPage, HowTo, and Product. Pick the one type that best describes the page and use it.

3. Heading-query mismatch. Your heading says "Meta Tag Best Practices" but users ask "how do I fix my meta tags?" The heading needs to match the conversational query, not the topical category.

4. No entity linking. Missing sameAs, missing @id, missing author references. Without these, the answer engine has no way to confirm you're a real publisher with a real track record.

5. Skipping dateModified. Freshness is a first-class AI citation signal in 2026. Undated content is treated as stale by default.

6. Wrong schema type. QAPage where FAQPage belongs. HowTo on conceptual content. Article on a product page. The type conveys intent — use the one that matches.

7. Schema without source attribution. FAQ answers that state claims with no citations to authoritative sources. AI assistants treat unsourced claims with increasing skepticism, and sourced answers get cited at higher rates.

▲ Overt Ops

Auditing every page of a real site for these seven failures, across FAQ, HowTo, Speakable, and Q&A schema, is the kind of work that eats a full engineering week. Overt Ops scores your AEO readiness automatically — entity clarity, factual density, schema validity, answer-block coverage — and generates the JSON-LD fixes in your voice, ready to paste. Coming soon from Area 51 Software.

Building Your AEO Rollout Plan

You don't need to retrofit the entire site in one sprint. Start where the leverage is highest:

Week One — The High-Value Pages

Week Two — The Tutorial Pages

Week Three — The Trust Layer

Week Four — Voice & Expansion

Take Action

The gap between "indexed" and "cited" is where the 2026 traffic opportunity lives. Most of your competitors have passable Core SEO and Open Graph coverage. Almost none of them have a real AEO layer. The window for being one of the first pages answer engines learn to cite for your queries is open now and closing fast.

Start with one page. Pick a high-intent, question-shaped query you want to own. Add the heading. Write the 40-60 word answer block. Add the FAQ schema. Validate. Paste the URL into Perplexity and watch what happens.

If it gets cited, repeat the pattern on the next nine pages. If it doesn't, the failure mode is almost always in the seven-failures list above — check invisible markup first, heading-query mismatch second, and missing entity signals third.

For full-site AEO auditing with generated JSON-LD fixes you can deploy immediately, keep an eye on Overt Ops.

OVERT OPS — COMING SOON

AI-powered SEO + AEO auditing. Full FAQ, HowTo, Speakable, and Q&A schema generation. Citation-readiness scoring across the four major answer engines. Actionable fixes in seconds.

Coming Soon