AEOcommercial intent

WordPress AI Visibility Checklist: 12 Fixes That Make Your Site Visible to ChatGPT, Claude, and Gemini

A practical WordPress guide covering robots.txt, llms.txt, JSON-LD schema, plugin conflicts, and theme pitfalls that block AI crawlers. Includes plugin-specific fixes for Yoast, Wordfence, and WooCommerce.

Feb 21, 202613 min readWordPress site owners, WooCommerce operators, and WordPress consultants
wordpress ai visibility optimizationhow to add llms.txt in wordpresswordpress robots txt for aiwordpress aeo checklistwordpress chatgpt visibilitywoocommerce ai seo

Why WordPress sites are disproportionately invisible to AI

WordPress powers 43% of the web, but a large share of WordPress sites are invisible to AI platforms. The reason is structural: WordPress relies heavily on plugins for SEO, security, and caching — and these plugins frequently generate robots.txt rules, meta tags, and security policies that block AI crawlers without the site owner knowing.

A typical WordPress site might have Yoast SEO generating its robots.txt, Wordfence adding bot-blocking firewall rules, and a caching plugin serving static HTML that strips JSON-LD structured data. Each layer independently seems fine, but the combined effect makes the site unreadable to GPTBot, ClaudeBot, and PerplexityBot.

The irony: WordPress site owners often produce excellent content that AI models would happily recommend — if they could access it. Fixing the technical blockers is usually a one-time setup that takes an afternoon.

Fix 1: Audit your robots.txt (plugins generate it automatically)

In WordPress, robots.txt is usually generated dynamically by your SEO plugin, not stored as a static file. Visit yourdomain.com/robots.txt and check what's actually being served.

If you're using Yoast SEO, go to Yoast > Tools > File Editor to see and edit the robots.txt output. If you're using All in One SEO, go to AIOSEO > Tools > Robots.txt Editor. The key check: make sure there's no 'Disallow: /' under 'User-agent: *' and no specific blocks for GPTBot or ClaudeBot.

Important: editing robots.txt directly via FTP can be overridden by your SEO plugin on the next update. Always make changes through the plugin's interface to ensure they persist.

Fix 2: Disable aggressive bot blocking in security plugins

Wordfence, iThemes Security (Solid Security), and Sucuri all have bot-blocking features that can misidentify AI crawlers as threats.

In Wordfence: go to Firewall > Blocking and check for any rules that block user agents containing 'bot' or 'crawler'. Also check Firewall > Rate Limiting — aggressive rate limits can throttle or block AI crawlers that request multiple pages. In iThemes Security: check Settings > Bot Fight Mode and ensure AI crawlers aren't being challenged.

If you can't find the specific setting, temporarily disable the security plugin's firewall (not the entire plugin) and re-test robots.txt access with a GPTBot user agent. If the block disappears, the security plugin is the culprit.

Fix 3: Create and host llms.txt

WordPress doesn't generate llms.txt by default. You need to create it manually and serve it at yourdomain.com/llms.txt.

The simplest approach for a standard WordPress site: create a file called llms.txt in your WordPress root directory (the same directory as wp-config.php) with your site name, a one-paragraph description, and links to your most important pages — homepage, top category pages, key product pages, about page, and contact page.

For WooCommerce stores: include your shop page, top product categories, shipping policy, returns policy, and 5-10 best-selling product pages. Avoid including every product URL — llms.txt is a priority signal, not a sitemap.

If your hosting doesn't allow file uploads to the root directory, you can create a WordPress page at /llms-txt and use a redirect rule in .htaccess to serve it at /llms.txt. Some WordPress hosting providers (like WP Engine) also allow adding files through their control panel.

Fix 4: Add JSON-LD structured data

Most WordPress SEO plugins add basic structured data, but the defaults are often incomplete or incorrect for AI visibility purposes.

Check what Yoast or AIOSEO generates by viewing your page source and searching for 'application/ld+json'. You should find at minimum an Organization schema on the homepage and Article or Product schema on content pages. If these are missing, your SEO plugin may have structured data disabled — re-enable it in the plugin settings.

For WooCommerce: verify that product pages have Product schema with price, availability, currency, and description fields. WooCommerce's default schema is often overridden by Yoast WooCommerce SEO — check which plugin is generating the schema and ensure the data matches what's visible on the page.

Add FAQPage schema to any page with Q&A content. This is high-impact because AI models disproportionately cite FAQ-formatted content. You can add this manually in your theme's functions.php or use a dedicated schema plugin like Schema Pro.

Fix 5: Ensure content renders without JavaScript

Some WordPress themes and page builders (Elementor, Divi, WPBakery) render significant portions of content via JavaScript. AI crawlers have inconsistent JavaScript support — GPTBot handles it well, but ClaudeBot and PerplexityBot may only see the initial HTML.

Test this: view your page source (not the inspected DOM) and check whether your key content (product descriptions, pricing, contact info) appears in the raw HTML. If it only appears after JavaScript runs, it may be invisible to some AI crawlers.

For page builder sites, use the builder's 'static HTML' or 'server-side render' options if available. For critical content like pricing tables or product specifications, consider using native WordPress blocks or shortcodes that render server-side.

Fix 6: Clean up canonical tags and redirects

WordPress generates canonical tags automatically, but plugin conflicts and permalink changes can create inconsistencies. A product page with a canonical tag pointing to a different URL confuses AI models about which URL is authoritative.

Check for these common WordPress canonical issues: HTTP vs HTTPS conflicts (canonical points to http:// while the site serves https://), www vs non-www mismatches, trailing slash inconsistencies, and paginated archive pages with canonicals pointing to page 1.

Also check for redirect chains. AI crawlers may follow redirects, but each hop reduces the confidence signal. A URL that redirects three times before reaching the final page is less likely to be cited than a direct URL.

WooCommerce-specific fixes

WooCommerce stores have additional AI visibility requirements beyond standard WordPress sites.

  • Product variations (size, color) should have distinct URLs with unique structured data, not just JavaScript-swapped content on the same URL. AI models can't interact with dropdown selectors.
  • Out-of-stock products should return 200 status codes with updated availability in their schema, not 404 or redirect to the category page. AI models need to know the product exists but is temporarily unavailable.
  • Category and tag archive pages should have unique descriptions and structured data, not just auto-generated lists. These pages often rank well and are frequently cited by AI models for discovery queries.
  • Shopping cart and checkout pages should be excluded from llms.txt and structured data — they're user-specific and provide no value for AI visibility.
  • Review data should be included in Product schema as aggregate ratings. AI models use review counts and ratings as quality signals when choosing which products to recommend.

Execution Checklist

  • Check robots.txt via your SEO plugin's interface (not direct file editing) — look for global blocks and bot-specific rules.
  • Audit security plugin settings: Wordfence firewall rules, iThemes bot fight mode, Sucuri bot blocking.
  • Create llms.txt in your WordPress root directory with your 10-20 most important pages.
  • Verify JSON-LD structured data on homepage (Organization), posts (Article), and products (Product).
  • Add FAQPage schema to any page with Q&A content.
  • Test that key content appears in raw HTML source, not just JavaScript-rendered DOM.
  • Check canonical tags for HTTP/HTTPS, www/non-www, and trailing slash consistency.
  • For WooCommerce: verify product variations have distinct URLs, out-of-stock products return 200, and review data is in Product schema.
  • Re-check after every plugin or theme update — WordPress regenerates robots.txt and schema dynamically.

FAQ

Is there one WordPress plugin that solves AI visibility?

No single plugin covers everything. SEO plugins (Yoast, AIOSEO) handle structured data and robots.txt. Security plugins (Wordfence) control bot access. And llms.txt currently requires manual setup. The closest all-in-one approach is using your SEO plugin for schema and robots.txt, whitelisting AI bots in your security plugin, and manually maintaining llms.txt.

Do blog-only WordPress sites need AI visibility optimization?

Yes. Blog content is frequently cited by AI models in informational responses. If someone asks ChatGPT a question your blog answers, you want to be the source it cites. The same fixes apply: allow AI crawlers, add Article schema, create llms.txt pointing to your best content, and ensure your content renders server-side.

Will these changes affect my Google Search rankings?

The changes described here are either neutral or positive for Google Search. Fixing canonical tags, adding structured data, and cleaning up redirects are SEO best practices. Allowing AI crawlers via robots.txt has no effect on Googlebot — they're separate user agents. The only consideration is Google-Extended, which controls Google's AI products (Gemini) but does not affect Google Search indexing.

Run Free AuditView PricingBack to Blog

Related Posts