Skip to main content

No AI-blocking meta robots directives

ab-001520 · geo-readiness.ai-crawler-access.no-ai-blocking-meta-tags
Severity: highactive

Why it matters

The noai and noimageai meta robots directives instruct large language models and image-generation systems to exclude your pages from their training corpora and retrieval indexes. Shipping these directives on public marketing pages silently removes your product from ChatGPT, Gemini, Perplexity, and Copilot answers — competitors without the tag get cited in your place, and prospects asking AI assistants for recommendations in your category will never hear your name.

Severity rationale

High because the directive fully excludes the page from AI citation surfaces with no partial-credit fallback and no visible warning.

Remediation

Remove the noai and noimageai values from every public page's robots metadata and from any X-Robots-Tag header set in middleware or server config. In Next.js App Router, audit each layout's metadata.robots export and scoped directives only to admin or user-generated-content routes if needed.

export const metadata = {
  robots: { index: true, follow: true },
}

Detection

  • ID: geo-readiness.ai-crawler-access.no-ai-blocking-meta-tags

  • Severity: high

  • What to look for: Search across all page layouts and page files for meta robots directives containing noai or noimageai. Count all layout files and page files that set meta robots directives. In Next.js App Router, check metadata exports for robots properties. Also check for X-Robots-Tag headers in middleware or server config that include AI-blocking directives. Before evaluating, extract and quote the exact robots metadata configuration from the root layout file, if present.

  • Pass criteria: Count all occurrences of noai and noimageai across all layout and page files. The count must be 0 on public content pages. Scoped directives on non-public pages (admin, user-generated content) are acceptable — pass with a note listing which non-public pages have the directives. Report even on pass: "Scanned X layout/page files — 0 noai/noimageai directives found on public pages."

  • Fail criteria: At least 1 blanket noai or noimageai directive found on public-facing content pages (homepage, marketing pages, blog, docs). Report: "Found noai/noimageai on X public pages: [list routes]".

  • Skip (N/A) when: Never — applies to all web projects.

  • Detail on fail: "Meta robots 'noai' directive found in root layout — blocks AI systems from processing all page content across all routes" or "X-Robots-Tag header includes 'noimageai' for all routes — 0 of N public pages are AI-accessible"

  • Remediation: The noai and noimageai meta robots directives tell AI systems not to use your content. Remove these from public content pages:

    // Remove noai from metadata
    export const metadata = {
      robots: {
        index: true,
        follow: true,
        // Do NOT include: noai: true, noimageai: true
      },
    }
    

Taxons

History