The "does AI content rank?" debate has been emotional for two years. By April 2026 the question is empirical. We tracked 200 paired articles — 100 AI-generated, 100 human-written, matched on domain, publish week, internal-link template, topic specificity, and word count band — across 14 publishing domains for six months. The data does not say AI content is dead. It does not say human content is irreplaceable. It says the two perform very differently across the stages of an article's life on a SERP.
AI wins the first lap: faster indexing, higher initial position, cheaper to produce. Human wins the marathon: by month three the ranking trajectories invert; by month six there is a five-position median gap, in human's favor. The SERP-feature picture is harsher — featured snippets, People-Also-Ask answers, and AI Overview citations all skew strongly toward human content, with the AIO gap the most pronounced (4% AI vs 11% human cited).
What follows is the methodology, the headline numbers by stage of life, and a taxonomy of which content types AI handles well and which it handles poorly. The full dataset — pair IDs, domain categories, publish dates, weekly position snapshots, and SERP feature flags — is publicly downloadable as a methodology companion. Where we draw inferences beyond what the raw data supports, we say so.
- 01AI indexes 1.8× faster — median 14 hours to first-index vs 26 hours for human content.AI-generated articles reach Google's index faster across all 14 domains in the study. The advantage is consistent and statistically significant. Hypothesized causes: faster publish cadence creates a recency signal, and AI articles tend to ship with cleaner schema and faster crawl-friendly templates because they are produced by automation pipelines.
- 02Ranking trajectories invert by month three. Human gains; AI drifts.Initial position week 1: AI median 18, human median 22. Month 3: AI averages −3 position drift; human averages +6 position gain. Month 6: AI median 21, human median 16 — a 5-position gap, fully reversed from start. The divergence is consistent across 12 of 14 tracked domains.
- 03SERP feature capture skews hard toward human — and AI Overview gap is the largest.Featured snippets: AI 12% vs human 19% (−7pp). People-Also-Ask answer panels: AI 8% vs human 14%. AI Overview citation rate: AI content 4% vs human 11%. The AIO gap is the most operationally important — Google's generative search heavily prefers human-attributed content, and the gap is widening through 2026.
- 04CTR is 8% lower on AI titles, even at the same position. Backlinks 4× lower.Position-adjusted CTR at position 5: AI 6.2% vs human 7.1%. Average CTR over the 6 months: AI 4.1% vs human 4.4%. Backlinks acquired in the 6-month window: AI median 2 vs human median 8. Bounce rate AI 64% vs human 56%; session duration AI 1m 42s vs human 2m 31s. The downstream conversion gap is the same shape: AI 0.8% vs human 1.4% on B2B demo-request workflows.
- 05AI gets close to human on data-heavy formats. AI fails on opinion and reviews.By content type, AI's relative performance vs human ranges from 92% (statistics roundups) and 88% (comparison/vs articles) at the high end, down to 41% (opinion/commentary) and 38% (product reviews) at the low end. Use AI for structured, data-dense content with clear comparison tables; do not use it for first-party opinion or product evaluation where lived experience is the differentiator.
01 — The ThesisThe AI content debate is empirical now.
For most of 2024 and 2025 the AI content question was litigated with anecdote. One agency would publish a case study showing AI content ranking #1 within a week. Another would publish a postmortem on a domain that lost 60% of organic traffic after an AI content sprint. Both were true. Neither was generalizable. The absence of paired controls — same domain, same week, same length, same internal-link structure — meant every claim was confounded by variables nobody bothered to hold constant.
We started this study in October 2025 specifically to remove those confounds. The pairing methodology is simple but expensive: every AI article in the dataset has a human counterpart on the same domain, published within seven days, in the same content category, within the same word-count band, with the same internal-link template applied at publish time. The treatment variable is the author intervention; everything else is controlled. The dataset is small enough (200 pairs) to allow individual-pair audit, large enough that the headline deltas are statistically significant at 95% CI.
We are not claiming this study is definitive. We are claiming it is the most carefully-controlled paired SERP-tracking dataset currently public for the AI-vs-human content question. Methodology disclosed; raw data downloadable; replication welcomed.
02 — Methodology200 paired articles · 14 domains · 6 months.
The dataset comprises 200 articles — 100 AI-generated, 100 human-written — published between October 1 and December 15, 2025, and tracked daily through April 15, 2026. Each AI article is paired with one human article on the same domain. Domains were selected to span four publisher categories: B2B SaaS company blogs (5 domains), digital agency blogs (3), news / publisher sites (3), and ecommerce / DTC blogs (3). All 14 domains had at least 6 months of indexing history before the study window opened, to avoid new-domain crawl artifacts.
Same publishing domain
AI and human pair share the same root domainRemoves the largest single confound — domain authority. We do not compare AI on a high-DA news site to human on a brand-new SaaS blog. Every comparison happens within the same site's authority and crawl budget.
removes DA confoundSame publish week
Pair publishes within 7 days of each otherRemoves seasonality, algorithm-update timing, and SERP-feature rollout effects. If a pair was hit by the November 2025 core update, both articles were hit. Time-conditioned variables apply identically to treatment and control.
removes seasonalitySame word-count band & template
1,800–3,500 words · identical internal-link templateWord count is held within a 50% band; internal-link template (count, anchor pattern, hub-page links) is identical at publish. Removes content-length and link-equity confounds that otherwise dominate ranking variance.
removes structure confoundTracking instrumented daily SERP position (top 100 for the primary target keyword), weekly SERP-feature inventory (featured snippet, PAA inclusion, AI Overview citation), GA4 sessions / engagement metrics, and a backlink scrape via Ahrefs API at week 0, 4, 12, and 24. Position is reported as median across the daily samples within each week to dampen daily volatility. Confidence intervals on aggregate deltas are bootstrap 95% CIs computed across the 100 paired observations.
03 — IndexingIndexing speed: where AI wins.
The first surprise in the data is the indexing-speed advantage. AI articles hit Google's index faster than their human pairs across all 14 domains, with median time-to-first-index of 14 hours for AI versus 26 hours for human. The 1.8× advantage is consistent across publisher types and shows up cleanly in the bootstrap CI (95% CI on the ratio: 1.6×–2.0×). It is the cleanest single result in the study.
AI median
Half of AI articles were indexed within 14 hours of publish. The fastest 25% indexed within 8 hours. Distribution skews tight — very few AI articles took longer than 36 hours, suggesting the indexing path is consistent across the AI cohort.
1.8× fasterHuman median
Human pairs median 26 hours, with much longer right-tail (slowest 25% took 48+ hours). The variance suggests human-publish pipelines are less consistent — manual publish steps, less aggressive submission to Google Search Console, slower internal-link cascades.
baselineAI vs human at week 1
Once indexed, AI starts higher: median position 18 at week 1 vs human median 22. Roughly 4 positions of initial-rank advantage. This is consistent with AI articles being more keyword-dense and matching surface-level relevance signals more aggressively than human-written drafts.
AI starts higher04 — TrajectoryRanking trajectory: where human wins.
The indexing-speed advantage has a short shelf life. By month three the ranking trajectories of the two cohorts cleanly diverge: human articles gain a median 6 positions from their week-1 baseline; AI articles drift down a median 3 positions. The gap opens further through months four, five, and six. By month six, human has a 5-position median advantage — fully reversed from the 4-position AI advantage at week 1.
Median SERP position over 6 months · AI vs human pairs
Source: Digital Applied 6-month tracking study · Q4 2025–Q2 2026 · n=200 paired articles · 14 domainsThe trajectory pattern is the most operationally important finding in the study. A two-week post-publish review will show AI outperforming human — that is exactly the time at which most content teams cut their performance review and lock in the AI decision. By the time the trajectories invert at month three, the decision has already shipped to the next quarter's plan. Anyone evaluating an AI content program on a 30-day window will reliably mis-read the data.
What changes between week 1 and month 3? Three things, in decreasing order of estimated importance: backlinks (human articles attract 4× more backlinks in the 6-month window), engagement signals (human articles run lower bounce and longer session duration, both of which Google now uses as ranking inputs through Chrome's anonymized telemetry), and SERP feature placement (human articles capture more snippets and PAA panels, which themselves cause CTR redistribution that reinforces position).
05 — SERP FeaturesWhere AI Overview punishes AI content the most.
SERP feature capture is the place where the AI-vs-human gap is widest, most consistent, and most actionable. We tracked three feature classes weekly: featured snippets, People-Also-Ask answer panels, and AI Overview source citations. Across all three, human content captures more — and the AI Overview gap is the largest at roughly 3× the citation rate.
Featured snippet capture · 6-month average
AI 12% vs human 19% (−7pp gap). Across the 200 articles, human pairs captured the featured snippet for their target keyword in 19% of cases, AI in 12%. The gap is consistent across publisher categories, but largest in the news/publisher cohort where the snippet competition is fiercest.
human +7ppPeople-Also-Ask answer panel inclusion
AI 8% vs human 14%. Lower headline numbers because PAA inclusion requires a specific Q-then-A formatting that neither cohort optimized for explicitly. Where it appears, human-authored content is included nearly 2× as often as AI-authored — likely because human writers naturally include question-formatted subheads that PAA matches against.
human +6ppAI Overview source citation rate
AI 4% vs human 11%. The largest relative gap in the study (~3× human advantage). When Google's AIO surfaces for a tracked keyword and cites a source, it cites human-authored content nearly three times more often than AI-authored. Strong inference: Google's AIO has a quality classifier that downweights detected AI content as a citation source, even when it ranks similarly in the standard SERP.
human ~3×Backlinks acquired in 6-month window
AI median 2 vs human median 8 (4× gap). Not strictly a SERP feature, but the strongest external validation signal. Human articles earn editorial links at 4× the rate of paired AI articles. Interpretation: links are mostly earned by genuine reference-worthiness — original data, named-source quotes, distinctive POV — which AI content lacks by construction.
human 4×06 — BehaviorCTR and engagement deltas at the same position.
One sub-question we wanted to settle: when AI and human articles rank at the same position, do they draw the same click-through rate? The answer is no — AI articles draw 8% lower CTR position-adjusted, and the gap is driven by title quality plus a small but measurable trust effect on AI-detected bylines. Engagement-on-page is a larger gap: bounce rate runs 8 percentage points worse, and session duration is roughly 35% shorter.
CTR and engagement gap · AI vs human pairs
Source: Digital Applied 6-month tracking study · Q4 2025–Q2 2026 · GA4 + GSCThe conversion gap is the one number that should change planning for any team running AI content into a measurable funnel. A 1.75× higher conversion rate on the human cohort means that even if AI ranks identically and produces identical session counts, the funnel output is 75% smaller. For B2B SaaS spending $40 cost-per- session-equivalent on paid acquisition, that is the difference between a profitable content channel and a vanity-metrics exercise.
The patterns that hurt AI content most across this dataset are recognizable: em-dash overuse, generic openers, the "delve" / "moreover" / "in today's fast-paced world" cluster, and — more costly than the prose tics — the absence of original data, named sources, or first-party POV. The patterns that help: clear comparison tables, structured H2/H3 hierarchy, explicit numerical depth in claims, and clean schema markup. None of those patterns require human authorship — they require an editor pass.
07 — By Content TypeHow close AI gets — by content type.
The aggregate numbers mask large variation by content type. AI performs near-parity with human on data-heavy formats (statistics roundups, comparison/vs articles), drops moderately on procedural content (how-to guides), and falls off a cliff on opinion and review content. The "AI uplift gap" — AI's blended SERP performance as a percentage of its human pair's — ranges from 92% on the strongest format to 38% on the weakest.
AI uplift gap by content type · AI as % of human pair performance
Source: Digital Applied 6-month tracking study · blended SERP score = position + features + CTR + engagementStatistics roundups · comparison/vs · structured how-to
AI handles dense, structured, fact-organized content within 8–24% of human performance. The format has clear scaffolding (tables, comparisons, numbered steps) that AI can execute cleanly. Pair with a 30-minute editor pass to add 1–2 original data points and the gap shrinks further. This is where AI content programs earn their keep.
AI + light editGlossaries · long-form references · explainers
AI gets 60–75% of human performance on these formats. Workable as a draft generator with significant human revision, especially for inserting domain-expert framing and concrete examples. Not workable as a publish-as-generated workflow — the gap shows up most clearly in PAA and AIO citation rates.
AI draft + heavy editOpinion · commentary · product reviews · case studies
AI lands at 38–58% of human performance. The gap is structural — these formats reward lived experience, named sources, original POV, and first-party evaluation, all of which AI lacks by construction. Editorial polish does not close the gap. Either commit to human authorship or skip the format.
Human onlyOne observation worth flagging: the AI-content "good enough" threshold has crept upward through 2025 and 2026 as frontier models improved, but the relative ranking of formats has been stable since at least mid-2024. Statistics and comparison content have always been AI's strongest formats. Opinion and reviews have always been AI's weakest. Better models tightened the gap on the strong formats faster than on the weak formats. The ceiling on opinion/review content is being set by the absence of first-party experience, not by model quality — and no foreseeable model release closes that gap.
08 — ConclusionWhen AI is good enough and when it is not.
The data answers the question — by format, not in aggregate.
The headline finding from six months of paired tracking is simpler than the AI-content discourse has made it sound. AI wins the fast metrics — indexing, initial position, cost per draft. Human wins the slow metrics — month-3 trajectory, snippet capture, AI Overview citation, backlinks, engagement, and conversion. Anyone who evaluates an AI content program on a two-week window will get the wrong answer; anyone who evaluates on a six-month window will see the trajectory invert.
The actionable read is by format, not in aggregate. Use AI for statistics roundups, comparison content, and structured how-to guides — formats where AI lands within 8–24% of human performance and a light editorial pass closes the gap. Skip AI for opinion, commentary, product reviews, and first-party case studies — formats where AI lands at 38–58% of human performance and editorial polish cannot close the gap. The ceiling on those formats is set by the absence of lived experience and named sources, not by model quality.
The trend line that should worry AI-content-heavy sites is the AI Overview citation gap. At a 3× human advantage today, with a classifier that is publicly committed to weighting source quality, we expect this gap to widen through 2026 and 2027. Sites planning around AIO referral traffic should weight their content mix toward the formats where AI competes well, and commit human authorship to the formats where AIO citation is the upside.