SEOJul 27, 2025·12 min read

The State of AI-Generated Content in Google: What the Data Says About Indexing and Ranking

Capconvert Team

Content Strategy

TL;DR

Every week, a new headline declares that AI content is either the future of SEO or the fastest way to get your site deindexed. The confusion is understandable. Google's messaging has evolved rapidly, multiple algorithm updates have reshaped the playing field, and the actual data from large-scale studies often contradicts the loudest voices in the room. Here's what's real: according to Ahrefs' analysis of 600,000 pages, 86.

Every week, a new headline declares that AI content is either the future of SEO or the fastest way to get your site deindexed. The confusion is understandable. Google's messaging has evolved rapidly, multiple algorithm updates have reshaped the playing field, and the actual data from large-scale studies often contradicts the loudest voices in the room. Here's what's real: according to Ahrefs' analysis of 600,000 pages, 86.5% of top-ranking pages contain some amount of AI-generated content. At the same time, the February 2026 core update sent Semrush Sensor to 9.4 as mass AI content sites saw 40–60% traffic drops. These two facts aren't contradictory. They're the clearest evidence yet that Google has drawn a precise line-and most SEOs are standing on the wrong side of the debate. This post breaks down what the data actually shows about AI content indexing, ranking performance, and the signals that separate content Google rewards from content it buries. No speculation. No vendor pitches. Just research, cross-referenced across the largest studies available.

Google's Official Position: What the Documentation Actually Says

Start with the primary source. Google's ranking systems aim to reward original, high-quality content that demonstrates qualities of E-E-A-T: expertise, experience, authoritativeness, and trustworthiness. Their focus on the quality of content, rather than how content is produced, has helped them deliver reliable, high-quality results to users for years.

That statement, published on Google Search Central in February 2023, remains their foundational policy. Google explicitly states that appropriate use of AI or automation is not against their guidelines, meaning it is not used to generate content primarily to manipulate search rankings.

The distinction is behavioral, not technical. Google recognizes that not all use of automation, including AI generation, is spam-automation has long been used to generate helpful content, such as sports scores, weather forecasts, and transcripts, and AI has the ability to power new levels of expression and creativity.

Where Google draws the line is on intent and quality. Using generative AI tools or other similar tools to generate many pages without adding value for users may violate Google's spam policy on scaled content abuse. That phrase-"scaled content abuse"-has become the operational boundary.

The January 2025 Quality Rater Guidelines Update

A significant shift arrived in January 2025. With its latest Search Quality Rater Guidelines update, Google added a definition and framing for generative AI for the first time. Google's document calls it a useful tool, but one that can be abused.

The key change was section 4.6.6. This new section targets main content created with little to no effort, originality, and added value. The Lowest rating applies if all or almost all of the main content on the page is "copied, paraphrased, embedded, auto or AI generated, or reposted from other sources with little to no effort, little to no originality, and little to no added value for visitors to the website."

Read that carefully. The trigger isn't AI origin. It's the absence of effort, originality, and value. According to Search Engine Land's analysis, raters are instructed to evaluate AI content on the same quality dimensions as human content. The guidelines rate AI content as "Lowest" quality only when it "lacks human oversight and review." Content that has been edited, fact-checked, and improved by humans doesn't receive this rating regardless of its origin.

The 600,000-Page Study: What Ahrefs Found About Ranking Correlation

The most cited dataset on this topic comes from Ahrefs, which analyzed 600,000 pages across 100,000 random keywords. The methodology was straightforward: they extracted the top 20 ranking URLs for 100,000 random keywords from Ahrefs Keywords Explorer, then ran each article through their own AI content detector.

The headline finding: the correlation between AI content percentage and search ranking position was 0.011-effectively zero. There is no clear relationship between how much AI-generated content a page has and how highly it ranks on Google. This suggests that Google neither significantly rewards nor penalizes pages just because they use AI.

But the nuance matters more than the headline. The very highest-ranking pages-those in position #1-tend to have slightly less AI-generated content. Pages with minimal AI use (0–30%) correlate very slightly with higher rankings, suggesting a minor preference for more human-created or lightly AI-assisted content at the very top of the SERPs.

Only 4.6% of top-ranking pages were fully AI-generated. The majority-81.9% of pages analyzed-offered a blend of AI and human input. That blend is the operating reality of modern content production. Pure AI content exists in search results, but it rarely dominates.

Where Semrush's Data Adds Context

Semrush conducted its own study, analyzing 20,000 blog URLs. They found that 57% of the AI content and 58% of the human content analyzed ended up in the top 10, meaning they have the same likelihood of ranking on page one.

However, the gap at the very top tells a different story. Content classified as purely AI-generated appeared in the top spot just 9% of the time, while content classified as human-written was there 80% of the time. Position 1 remains a human-dominated zone.

The only real separation is at the very top. From position 5 onward, the gap between human-written and AI-generated content is relatively narrow, so if most teams are benchmarking against "ranking on page one," AI content is holding its own.

The 16-Month Experiment: What Happens When AI Content Meets the Real World

A particularly revealing experiment tracked 2,000 AI-generated articles published across 20 brand-new domains over 16 months. The results followed a predictable arc that every SEO practitioner should understand. Phase 1 - Initial indexing was strong. About 71% of new AI-generated pages were indexed within the first 36 days. They generated over 122,000 impressions and 244 clicks. Even at this early stage, 80% of sites ranked for at least 100 keywords each.

Phase 2 - Early growth looked promising. Cumulative impressions grew to over 526,000, with 782 clicks. Content continued to perform well without backlinks, promotion, internal linking, or additional SEO tactics.

Phase 3 - The cliff arrived. By about three months, only 3% of pages remained in the top 100. Early relevance helped pages get indexed and briefly appear in search, but without authority, uniqueness, or E-E-A-T signals, rankings dropped sharply.

Phase 4 - Long-term stagnation. After over a year, visibility remained low across most sites. Impressions and clicks were minimal, and no site showed meaningful recovery.

The takeaway from this experiment isn't that AI content fails. The results don't mean AI content is useless-they show AI alone isn't enough to drive lasting impact. Early traffic and impressions may look promising, but without a clear SEO strategy and human guidance, those gains will likely fade within a few months.

This is the trap. Those first 36 days of indexing and impressions give teams false confidence. Early impressions often reflect testing rather than durable trust. Google may give new pages visibility to evaluate relevance and usefulness. That initial exposure should not be confused with proven ranking strength.

Originality.ai's Ongoing Study: How Much AI Content Is Actually in Google?

Originality.ai has been tracking AI content saturation in Google's top 20 results since 2019, sampling 500 popular keywords monthly. Their longitudinal data provides the clearest picture of AI's growing footprint.

Overall, from 2019 to 2025, AI content in Google search results has significantly increased. In February 2019, just 2.27% of the top 20 Google search results were AI. In July 2025, AI in Google skyrocketed to 19.56%. Yet as of September 2025, these levels have dipped to 17.31%.

The fluctuations correlate directly with algorithm updates. There are fluctuations in AI search ranking trends throughout the study period, such as with the March 2024 update, when AI levels in search rankings dropped to 7.43% from a previous high of 8.48% in December 2023.

The March 2024 core update was particularly aggressive. Google emphasized its goal of reducing unhelpful, irrelevant, unoriginal content from search results, promising to remove up to 40% of low-quality websites that provide useless information and poor user experience. Sites built entirely on scaled AI content were hit hardest. Yet AI content rebounded within months. That rebound tells us something important: Google's algorithms aren't detecting and penalizing AI origin. They're measuring quality signals that poorly produced AI content tends to lack-and that well-produced AI content can satisfy.

What Separates AI Content That Ranks From AI Content That Disappears

The data across these studies converges on a clear set of differentiators. Google doesn't evaluate a binary variable called "AI." It is evaluating pages and sites in context.

Authority and Trust Signals

The 16-month experiment failed partly because the domains had zero authority. The 2,000 articles lacked many signals Google uses to assess quality and trust: authority-no backlinks or external validation; and expertise and credibility-no authors, credentials, or real-world expertise.

Compare that with SE Ranking's parallel experiment, where AI-assisted articles published on their established blog performed differently. Six AI-assisted articles received almost 555K impressions and 2,300+ clicks between June 2024 and July 2025. Three of the six AI articles currently rank in the organic top 10. Five trigger AI Overviews and four are featured as sources. Domain authority made the difference.

Human Editorial Enhancement

Optimizing and enhancing AI content is key: 69% of marketers refine AI drafts with human editing, 48% build on initial AI drafts, and 55% conduct original research to strengthen their content.

This isn't optional polish. It's the mechanism that transforms generic AI output into content that satisfies E-E-A-T. While AI can mimic expertise by summarizing existing information, it cannot inject true personal experience or build genuine authority. Human editors, especially Subject Matter Experts, must review AI drafts to add unique insights, original case studies, and first-hand knowledge.

Content Differentiation

The experiment also lacked content differentiation. Much of the content resembled what already exists. Without unique insights, pages struggle to stand out. This is the Achilles' heel of default AI output. Language models are probabilistic summarizers-they synthesize existing information into statistically likely arrangements of words. Without deliberate injection of original data, proprietary perspectives, or practitioner experience, the output is inherently derivative.

SynthID and the Detection Question: Can Google Actually Tell?

A common anxiety among content teams: can Google detect AI content and penalize it algorithmically? The honest answer is more nuanced than either camp admits.

Google is building SynthID Watermarking-a proactive detection technology that marks AI-generated text, images, audio, and video with an invisible, machine-readable watermark directly upon creation. As of March 2026, over 10 billion pieces of content carry a SynthID watermark.

While originally focused on AI-generated imagery only, Google has since expanded SynthID to cover AI-generated text, audio and video content, including content generated by their Gemini, Imagen, Lyria and Veo models across Google. The infrastructure exists. Whether Google has formally wired SynthID into its search ranking algorithm hasn't been officially confirmed-but the infrastructure is sitting there, ready to go.

There are meaningful limitations. SynthID's most reliable detection is achieved with content generated by Google's own models. While an open-source version is available, detection rates and robustness are generally lower compared to Google's native implementations. SynthID's universal applicability remains limited, and it may not consistently identify AI-generated content from other providers such as OpenAI or Meta.

The practical implication: Google's ability to watermark and detect its own AI output is strong. Its ability to detect content from ChatGPT, Claude, or other models through SynthID alone is limited. But that may not matter. The correlation of 0.011 between AI content percentage and ranking position means Google isn't using detection as a ranking signal-it's using quality evaluation instead.

The Practitioner's Framework: How to Use AI Content Without Losing Rankings

The data points to a consistent operating model. AI is a production tool. Quality is a ranking signal. The two are separate concerns. Start with expertise, not prompts. The content that ranks at position #1 doesn't start with "write me a blog post about X." It starts with subject matter knowledge, proprietary data, or direct experience that gets structured using AI as an accelerator. Use AI to move faster through research, outlining, and drafting. Then invest the time saved into incorporating expert insights, proprietary data, and other elements that make your content stand out.

Build on established domains. The 16-month experiment proved that AI content on zero-authority domains craters after three months. The SE Ranking blog experiment proved that the same AI-assisted approach on an established domain generates hundreds of thousands of impressions. Your domain's existing authority isn't just a nice-to-have-it's the foundation that determines whether AI content gets a sustained hearing from Google's algorithms. Implement editorial quality loops. The Quality Rater Guidelines rate AI content as "Lowest" quality only when it "lacks human oversight and review." Content that has been edited, fact-checked, and improved by humans doesn't receive this rating regardless of its origin. Fact-checking, source verification, and editorial review aren't bureaucratic overhead. They're the specific signals that separate content Google's raters flag from content they approve. Avoid the scaled content trap. A company publishing 100 AI-generated pages that all follow the same template, contain no original information, and exist only to capture long-tail keywords is engaging in scaled content abuse. A company using AI to help produce well-researched, thoroughly edited articles that genuinely help their audience is just doing content production. The difference isn't volume-it's whether each page independently justifies its existence. Monitor for AI content patterns raters flag. Key signals that might trigger low-quality ratings include content with noticeable AI artifacts (phrases like "As an AI assistant…"), AI-generated summaries lacking accuracy or original insights, content that mimics human writing but provides no unique value, and text with unnatural repetition or phrasing patterns. Run your content through this lens before publication. Track the right metrics. Page-one ranking isn't the only indicator. Combining human writers with AI content led to a moderate improvement in conversions for 52% of users in Semrush's survey. Measure engagement, conversion rates, and time-on-page alongside ranking position. A page ranking #7 with high engagement and conversions may outperform a #3 ranking page with poor user signals.

Where This Is Heading: AI Overviews and the Shifting Definition of "Ranking"

The entire notion of ranking is being redefined. AI Overviews now reduce clicks by 58% according to Ahrefs' February 2026 data.

Around 93% of AI Mode searches end without a click-more than twice the rate of AI Overviews, where 43% result in zero clicks.

Being cited in AI-generated answers is becoming as important as ranking in traditional results. 76.1% of URLs cited in AI Overviews also rank in the top 10 of Google search results -meaning traditional SEO fundamentals still drive AI visibility. But the platforms diverge: in a dataset of 15,000 prompts, only 12% of links cited by ChatGPT, Gemini, and Copilot appear in Google's top 10 results for the same prompt.

Content freshness is also emerging as a differentiator. 85% of AI Overview citations were published in the last two years, and 44% are from 2025 alone. Stale AI-generated content ages faster than human-written content because it lacks the specificity and timeliness that AI systems increasingly favor when selecting sources. --- The data tells a coherent story. Google doesn't penalize AI content for being AI content. It penalizes content-regardless of origin-that lacks expertise, originality, authority, and genuine value to readers. The 0.011 correlation from the Ahrefs study confirms the mechanism. The 16-month experiment confirms the failure mode. The Quality Rater Guidelines confirm the evaluation framework. For practitioners, the path forward isn't about hiding AI usage or avoiding it entirely. It's about using AI as a production layer while ensuring every published page meets the same standard you'd apply to work from your best human writer: verified facts, original perspective, editorial polish, and clear value that justifies a reader's attention. The teams winning in this environment aren't the ones publishing the most content or the ones refusing to use AI. They're the ones who've built systems where AI handles the scaffolding while human expertise supplies the substance. That distinction-between AI as author and AI as tool-is the only one Google has ever cared about.

Ready to optimize for the AI era?

Get a free AEO audit and discover how your brand shows up in AI-powered search.

Get Your Free Audit