Google killed FAQ rich results for most websites in August 2023. Many teams took that as a signal to stop investing in FAQPage schema altogether. That was a mistake-but not for the reasons the SEO echo chamber typically offers. The question has shifted. It's no longer about earning an accordion dropdown in the SERPs. It's about whether the AI systems generating answers-ChatGPT, Google AI Overviews, Perplexity, Microsoft Copilot-actually read, process, and prioritize your JSON-LD when deciding which content to cite. The question isn't whether FAQ schema earns rich results anymore. It's whether it influences AI citation decisions, and through which mechanism.
The honest answer is more complicated-and more useful-than either "yes, implement everything" or "schema is dead." The evidence points to a dual-layer system where FAQ schema helps indirectly but powerfully, and where poor implementation can actively hurt you. This post breaks down what the data actually shows, what practitioners get wrong, and exactly where your structured data investment should go.
What Happened to FAQ Schema: The August 2023 Pivot
Google reduced the visibility of FAQ rich results and limited How-To rich results to desktop devices, then extended that deprecation to desktop as well, all in the name of providing "a cleaner and more consistent search experience."
Starting August 2023, FAQ rich results would only be shown for well-known, authoritative government and health websites, with all other sites effectively locked out.
The traffic impact was immediate. Before this restriction, FAQ rich results occupied up to four rows in SERPs and drove meaningful CTR increases. 65% of Schema App clients experienced click drops during an earlier FAQ fluctuation in April 2022. When Google pulled the plug for most domains, teams that had built entire content strategies around FAQ dropdowns saw their investment evaporate overnight. But here's what many teams missed: Schema App asked a deeper question-what is Schema Markup actually for? The answer was never rich results. It was always about helping machines understand meaning. That distinction became the foundation for a completely different strategic rationale-one built around AI citation rather than SERP decoration.
How LLMs Actually Process Your Structured Data
This is where the conversation gets honest, and where most published guides fall short. The mechanics of how AI systems interact with schema are layered, platform-specific, and far from settled.
The Discovery vs. Retrieval Distinction
The process splits into two phases: discovery and retrieval. Discovery is finding candidate URLs, where traditional search engines like Bing still do the heavy lifting, narrowing the entire web to a manageable set of pages. Retrieval is what happens next-the system fetches those pages, breaks them into chunks, and figures out which pieces answer the user's question.
Schema plays different roles at each stage. During discovery, your structured data enriches search engine indexes-especially Bing's, which feeds ChatGPT and Copilot. ChatGPT does not semantically parse JSON-LD-it treats schema as text during retrieval. The benefit is indirect: schema enriches Bing's index, and since ChatGPT uses Bing for its search-grounded responses, well-implemented schema improves how your content is understood and surfaced.
The SearchVIU Experiment: What AI Actually Sees
SearchVIU conducted a practical test with a specially developed test page for a fictional product. The goal was to determine whether popular AI systems can extract information from various sources on a web page, including schema-only data.
The results were sobering. SearchVIU tested eight different scenarios across five AI systems. In one test, they placed a product price exclusively in JSON-LD schema, not visible anywhere on the page. The result: zero out of five systems extracted it. They also tested hidden Microdata and hidden RDFa with the same result.
This tells us something important: if schema is used in the process at all, it is probably not being read the same way as visible content on the page. But the AISO experiment offered a critical counterpoint. AISO created two identical websites about a fictional company. Same content, same design, same visible information. The only difference: one site had comprehensive schema markup, one didn't.
Without schema, ChatGPT returned basic pricing info but missed the ratings entirely, even though they were visible on the page. With schema, ChatGPT returned pricing plus user ratings and review counts.
The conclusion that emerges: schema affects what gets noticed and extracted, even when the underlying content is identical.
The Microsoft Confirmation
The biggest confirmation came in March 2025 when Fabrice Canel, Principal Product Manager at Microsoft Bing, confirmed that schema markup helps Microsoft's LLMs understand content during his presentation at SMX Munich. Microsoft uses structured data to support how its large language models interpret web content, specifically for Bing's Copilot AI.
In March 2025, both Google and Microsoft publicly stated that they use Schema Markup for their generative AI features. Google was explicit: structured data is critical for modern search features because it is efficient, precise, and easy for machines to process.
Meanwhile, there are no peer-reviewed studies on schema's impact on AI search visibility. OpenAI, Anthropic, Perplexity, and other platforms besides Microsoft or Google haven't published their indexing methods. Practitioners who claim certainty about how every platform uses schema are extrapolating beyond the evidence.
The Data That Challenges the FAQ Schema Hype
Not all evidence supports the "FAQ schema supercharges AI citations" narrative. The SE Ranking study deserves particular attention because of its scale and methodological rigor.
SE Ranking analyzed 129,000 domains across 216,524 pages in 20 niches to identify which factors correlate with ChatGPT citations. Their finding on FAQ schema was counterintuitive: pages with FAQ schema averaged 3.6 citations, while pages without averaged 4.2.
The presence of FAQ sections within the main content nearly doubles your chances of being cited by ChatGPT. Yet, treat FAQ schema markup as optional, not essential. SE Ranking's data shows pages with FAQ schema average 3.6 citations, while those without reach 4.2. Schema alone does not significantly increase ChatGPT citation likelihood.
Read that distinction carefully. Visible FAQ content on the page helps enormously. The JSON-LD markup for that content may not add measurable citation lift for ChatGPT specifically. The Growth Marshal study added another dimension of nuance. A 2026 empirical study of 730 AI citations across ChatGPT and Gemini found that generic schema (Article, Organization, BreadcrumbList) provides zero measurable citation advantage. Only attribute-rich schema (Product and Review types with populated pricing, ratings, and specifications) showed a significant effect, cited at 61.7% versus 41.6% for generic implementations.
That's a critical finding. Generic, minimally populated schema actually underperforms having no schema at all-41.6% vs 59.8%. The CMS-default schema that most sites ship with isn't just ineffective. It may be actively working against you.
Why FAQ Schema Still Matters (But Not How You Think)
The conflicting data resolves when you understand that FAQ schema operates through two distinct pathways, not one.
Pathway 1: The Knowledge Graph Pipeline
When Google's crawler processes valid FAQPage JSON-LD, it extracts entity relationships and topical signals that feed Google's Knowledge Graph. This understanding influences organic rankings. Since 76% of AI Overview citations come from top-10 organic results, stronger Knowledge Graph representation leads to better organic rankings, which leads to higher AI Overview citation probability.
AI Mode shows expected results: pages with FAQ schema receive 4.9 citations versus 4.4 without it. The lift is modest but real for Google's own AI features, where the Knowledge Graph pipeline creates a direct line from schema to citation.
Pathway 2: Visible Content That Mirrors Schema
FAQ schema indirectly improves AI citation probability through Google's Knowledge Graph pipeline, and visible on-page Q&A content (which mirrors the schema) is directly extractable by every major AI platform. The most effective approach combines both layers: JSON-LD for Google's infrastructure, visible Q&A formatting for LLM extraction.
This is the practical takeaway most guides bury. LLMs don't parse your JSON-LD the way a search engine's structured data parser does. Both ChatGPT and Perplexity extracted data from invalid, made-up schema, proving they read <script> blocks as plain text, not as structured data. What they do respond to is well-organized, visible question-answer content that happens to also be marked up. The schema acts as a reinforcement signal-a parallel confirmation layer. When your visible content says "What is X?" followed by a concise answer, and your JSON-LD says the same thing in machine-readable format, you're creating redundancy that benefits both knowledge graph indexing and direct LLM extraction.
Platform-by-Platform: Where Schema Actually Moves the Needle
Treating "AI search" as a single channel leads to misallocated effort. Each platform behaves differently. Google AI Overviews benefit most directly from FAQ schema. 65% of pages cited by Google AI Mode include structured data. For ChatGPT, that number is 71%. Google's system has direct access to its own Knowledge Graph, making the schema-to-citation pipeline shortest and most reliable here. Microsoft Copilot is the platform with the clearest official endorsement. Microsoft is the only major platform to officially confirm schema helps its LLMs. Since ChatGPT and Copilot both use Bing's index, this is directly relevant.
ChatGPT presents the most complex picture. 28% of ChatGPT's most-cited pages have zero Google organic visibility. ChatGPT draws from an entirely different content pool than Google AI Overviews. A page that dominates AI Overviews may be invisible to ChatGPT, and vice versa. For ChatGPT specifically, the number of referring domains ranked as the single strongest predictor of citation likelihood. Backlinks, traffic, and trust scores ranked highest -schema sits further down the priority stack. Perplexity remains largely opaque. No public statement on schema from the platform. OpenAI's crawlers do not execute JavaScript. SearchVIU's experiment found Perplexity's bot found only 12.5% of test data points.
The implication is clear: if you're optimizing for Google AI Overviews, FAQ schema provides meaningful lift. If your primary target is ChatGPT, domain authority and content quality dwarf the schema signal.
The Attribute-Richness Rule: Why Half-Implemented Schema Backfires
The Growth Marshal finding deserves its own section because it fundamentally changes the implementation calculus.
The advantage is most pronounced for lower-authority domains (DR ≤ 60): attribute-rich schema achieves a 54.2% citation rate versus 31.8% for generic-a meaningful lift. Among high-authority domains (DR > 75), the gap narrows considerably.
This means the value of schema implementation follows a non-obvious pattern:
- Low-authority sites benefit most from comprehensive, attribute-rich schema-it becomes a competitive differentiator
- High-authority sites get cited regardless, making schema less impactful but still valuable as infrastructure
- Any site with thin, generic CMS-default schema is worse off than having no schema at all
For FAQ schema specifically, the attribute-richness principle means every question needs a genuinely useful answer, every answer should be 50–300 words (aligning with LLM retrieval chunk sizes), and the markup must exactly mirror visible page content. 3–5 questions per page, with each answer being 50–300 words in 2–4 self-contained sentences. This range aligns with LLM vector chunk sizes of 150–300 words, ensuring each answer fits within a single retrieval chunk without being split.
Don't mark up promotional content as FAQ. Don't add ten generic questions to pad out a page. Adding schema but neglecting content quality is a common mistake. Schema markup makes your content easier for AI systems to read. It doesn't make bad content worth citing. If the content itself doesn't follow AI citation best practices, schema alone won't save it.
The Agentic Web: Why Schema Investment Compounds Over Time
Beyond current AI search, there's a structural reason to invest in schema that most tactical guides ignore entirely.
Microsoft's NLWeb leverages semi-structured formats like Schema.org, RSS, and other data that websites already publish, combining them with LLM-powered tools to create natural language interfaces usable by both humans and AI agents. Microsoft believes NLWeb can play a similar role to HTML in the emerging agentic web.
The technical requirements confirm that a high-quality schema.org implementation is the primary key to entry. The NLWeb toolkit begins by crawling the site and extracting the schema markup. JSON-LD format is the preferred and most effective input for the system, consuming every detail, relationship, and property defined in your schema.
This isn't a theoretical future. NLWeb was built by the creator of Schema.org. It turns your existing structured data into a conversational interface for AI agents. Every NLWeb instance is automatically an MCP server.
Early NLWeb adopters include Shopify, Allrecipes, and Tripadvisor -companies making bets on a world where AI agents query websites programmatically rather than scraping them.
As AI systems began moving from answering questions to taking action, structured data became the connective tissue between websites and emerging agentic experiences. The schema you implement for FAQ citations today becomes the queryable interface that AI agents use tomorrow.
Implementation That Actually Works: A Practitioner Framework
Based on the evidence, here's how to approach FAQ schema for AI visibility in 2026. Start with your content, not your markup. Write genuine FAQ sections that answer real questions your audience asks. Use natural question phrasing as H2 or H3 headings. Lead each answer with a direct, 40–60 word response that can stand alone. Content structure and depth are equally important as technical signals. Longer articles, FAQ or Q&A sections, and question-based titles and headers all correlate with higher citation likelihood.
Layer schema types strategically. Pages with 3–4 complementary schema types (like Article + FAQPage + BreadcrumbList) get cited roughly 2x more often than pages with just one schema type.
Nesting FAQPage inside an Article schema creates a compound signal that tells AI engines both the content type and the specific Q&A pairs it contains.
Use JSON-LD exclusively. JSON-LD is the only real option for modern AI search optimization-it keeps markup separate from content, making it easier for AI crawlers to parse. Microdata and RDFa embed schema inside HTML tags, creating parsing conflicts when AI engines process rich text. JSON-LD lives in a dedicated script block, giving AI systems a clean, unambiguous signal layer.
Populate every relevant attribute. Don't ship minimal schema. Fill in author details, datePublished, dateModified, and connect entities with @id and sameAs properties. The lesson isn't "add schema." It's "add complete, accurate schema that faithfully mirrors visible page content."
Validate before deploying. Run every implementation through Google's Rich Results Test and Schema.org Validator. Every schema type you add should reflect content that's actually on the page. Adding FAQPage schema to a page with no visible FAQ section violates Google's structured data guidelines and can result in manual actions.
Measure with platform-specific tracking. After adding schema, check metrics at 30, 60, and 90 days. Schema changes don't take effect instantly. Track at the page level, not the site level. A site-wide average will dilute the signal. Use tools like Bing's AI Performance Dashboard, Semrush's AI Toolkit, and direct manual queries across ChatGPT and Perplexity to measure actual citation changes.
The Honest Bottom Line
FAQ schema won't rescue weak content. It won't compensate for poor domain authority. Schema carries approximately 10% weight in ChatGPT's citation evaluation, yielding a 3.5:1 authority-to-schema weighting ratio. FAQ schema cannot overcome weak domain authority, thin content, or low content quality. It's a last-mile optimizer for sites that already have the fundamentals.
But dismissing schema entirely is equally shortsighted. Schema markup is infrastructure, not a magic bullet. It won't necessarily get you cited more, but it's one of the few things you can control that platforms such as Bing and Google AI Overviews explicitly use.
The practical wisdom is this: if your content is strong, your authority is real, and your domain already earns organic visibility, well-implemented FAQ schema becomes the structural advantage that tips marginal citation decisions in your favor. It feeds Knowledge Graphs. It improves extraction accuracy. It prepares your site for the agentic web. And it compounds-every schema improvement you make today extends your machine-readable footprint into systems that haven't been built yet. The teams that win in AI search aren't the ones asking whether schema "works." They're the ones asking how to make their entire site readable, citable, and queryable by any system that comes along-and building that infrastructure before their competitors understand why it matters.
Ready to optimize for the AI era?
Get a free AEO audit and discover how your brand shows up in AI-powered search.
Get Your Free Audit