Skip to content
← ALL WRITING

2026-04-23 / 7 MIN READ

AI-assisted content with real grounding: the writer agent stack

Field notes on what real grounding looks like inside an AI writer agent pipeline, including the five grounding levels and where most AI content fails the test.

The word "grounding" gets used loosely in AI content conversations, usually to mean "the model has some reference data it pulled from." That is not what I mean by it, and it is probably not what Google's quality raters mean either when they distinguish helpful from unhelpful content. Grounding in the sense that matters is the chain of evidence from a specific real-world observation to a specific claim on the page.

These are field notes from working with DTC operators on content programs that publish at high cadence with a small team. This is what the grounding layer actually looks like in that kind of program, what makes it work, and where most content at scale fails the test.

Grounding level/Domain expertise
  1. 1
    General knowledge
    Paraphrasing common sources. Lowest utility.
  2. 2
    Public data
    Public datasets, industry reports. Neutral.
  3. 3
    Domain expertise
    Framework from documented practice.
  4. 4
    Case study
    Specific engagement, anonymized details.
  5. 5
    Original field data
    First-party measurements, unpublished.
Grounding strength scales the utility of AI-written pages. Level 3 and up hold up under HCU scrutiny.

What grounding actually means

A claim is grounded if a reader could, in principle, trace it back to a specific source that is not the page itself. Grounded claims have a provenance chain. Ungrounded claims are floating assertions.

Five grounding levels, from weakest to strongest.

Level 1: general knowledge. "Image optimization matters for SEO." This is a paraphrase of what every other page on the topic says. No provenance. Ungrounded.

Level 2: public data. "According to Google's Search Central documentation, INP replaced FID in March 2024." The claim is cited to a publicly-available source. Grounded but weak, because every other page can make the same citation.

Level 3: domain expertise. "The five requirements that cause programmatic pages to survive HCU rounds are (a), (b), (c), (d), (e), and here is the reasoning for each." The claim is framed as a synthesis of practice. Grounded because it is the author's framework; strengthens as the author's authority strengthens.

Level 4: case study. "I audited 14 Shopify stores on CAPI implementation and 8 of them were losing 20%+ of purchase events." The claim is grounded in a specific engagement, with enough detail to be credible. This is where AI-assisted content starts to become hard to distinguish from human bespoke content, because the grounding is real.

Level 5: original field data. "Here is the dataset from 140 DTC stores I benchmarked, the methodology, and the findings." This is the strongest form. It is also the form AI cannot generate on its own. A human has to do the benchmark.

Most AI content in 2022-2024 operated at levels 1 and 2. Most AI content that is working in 2026 operates at levels 3 and above.

The writer stack I recommend

A content program that produces grounded material at scale has five components.

One, the context library. A set of markdown files documenting the author's professional history, technical capabilities, case studies (anonymized), positioning research, and opinions on specific topics. About 15,000 words is a reasonable starting baseline. These become the reference library the drafting process leans on.

Two, the pipeline. A structured workflow that grounds every claim in the context library, cites real experience where possible, and flags any claim that cannot be grounded to a source. The pipeline should refuse to ship an unsourced strong claim.

Three, the voice calibration. Explicit voice rules (no em dashes, no staccato negation lists, first-person when grounded, humble register) plus samples of the author's writing from existing pieces. Without this, voice drifts across articles and a reader can feel the inconsistency.

Four, the human review step. Every draft gets read by the author. They fix factual errors, strengthen weak grounding, and rewrite any paragraph that reads as filler. The review is typically 10-20 minutes per article. That step is what converts a draft into authored work.

Five, the registry and the interlinking graph. Every article registers itself in a shared file. New articles link to siblings using varied anchor text. The cluster topology is explicit. This prevents the "each article is an island" failure mode.

What a grounded prompt looks like in practice

A grounded prompt for a piece on a narrow topic would include:

  • The cluster hub draft so the writer understands the broader argument
  • Two or three sibling articles so the writer knows which anchors are claimed
  • The author's professional context documents (call it a 15,000-word set)
  • A list of real grounding examples the author wants referenced
  • Voice rules and banned tokens
  • Frontmatter requirements

What the prompt should not include: "write 1,800 words on AI grounding." What it should include: "write this specific argument, using these specific real experiences, in this voice, at this length." The grounding is baked into the prompt, not expected to emerge from general knowledge.

Where AI content fails the grounding test

Three failure modes I see across client audits where the client has experimented with AI content.

One, generic prompts producing generic output. "Write a 1,500-word article about the benefits of programmatic SEO" produces level-1 output. The model paraphrases general knowledge. No grounding, no stance, no original data.

Two, hallucinated specifics. When you ask a model to be specific without giving it real specifics, it sometimes invents plausible-sounding statistics, company names, or case studies that do not exist. This is dangerous because it looks grounded but is not, and readers will occasionally catch it and lose trust.

Three, no review step. AI draft goes straight to publish. The review step is where the author validates the grounding. Skipping it means ungrounded claims ship alongside grounded ones, and the whole page reads as suspect.

The fix for all three is the same: treat the agent as a drafter, not a publisher. The human review step is non-negotiable.

How to audit your own AI content for grounding

Read through a page you published recently and score each paragraph on the 1-5 scale above. A grounded page has most paragraphs at level 3 or above. An ungrounded page has most paragraphs at level 1 or 2.

The pages that are going to survive long-term (and rank, and build author trust) are the ones where the grounding is real and traceable. The pages that are going to get demoted in a future HCU round are the ones where the grounding is vague or fabricated.

If you cannot defend the grounding for a specific paragraph, either rewrite it with real grounding or delete it.

The difference between ungrounded AI content and grounded AI content is mostly the prompt. The second non-negotiable is a human review step.

What this means for programmatic SEO at scale

A stack shaped like the one above can produce multiple articles per day. The grounding holds when the context documents are real, the voice is consistent, and the review step catches the failures.

That is a shippable pattern for a solo operator. It is also shippable for a small team. It is probably not shippable for a team trying to produce 50 articles per day across 20 unrelated topics, because the grounding breaks down when the author does not have real experience with every topic.

The constraint is not "AI can only write so much." The constraint is "a human can only ground so much." The right shape for scaled AI content is: narrow topic, real authority, human review, voice consistency.

Where this fits

These field notes sit alongside the cluster hub as the implementation detail for how to scale without losing utility. Programmatic SEO without the HCU demotion risk is the principled argument for why grounding matters. Content velocity with AI agents covers the velocity side of the same workflow.

For the broader agent-engineering perspective, the Claude Code agent handbook cluster covers how to build the kind of writer agents this workflow relies on.

If you want to set up this kind of grounded AI content stack on your own brand, the DTC stack audit includes content infrastructure review. Full product ladder is at /products.

Sources and further reading

  • Google Search Central: Helpful Content System documentation
  • Anthropic and OpenAI documentation on grounding and retrieval-augmented generation
  • Field notes from publishing AI-assisted content through 2023-2024 HCU rounds

// related

Claude Code Skills Pack

If you want to go deeper on agentic builds, this pack covers the patterns I use every day. File ownership, parallel agents, tool contracts.

>View the pack