4 pointsby aggeeinn2 hours ago2 comments
  • aggeeinn2 hours ago
    OP here.

    I’ve been trying to map out why some sites get cited by Perplexity/ChatGPT and others don't, so I built a custom crawler to audit 1,500 active websites (mix of e-commerce and SaaS).

    The most interesting findings:

    The Accidental Blockade: ~30% of sites are blocking GPTBot via legacy robots.txt rules or old security plugins (often without the owner knowing).

    The "Ghost Town": Only 3 sites (0.2%) had a valid llms.txt file.

    The JS Trap: 40% of marketing sites rely so heavily on client-side rendering that they appear as "empty shells" to non-hydrating AI agents.

    Context on the tool: I gathered this data using the engine for my project, Website AI Score. We are still in early beta (rough edges included), but we are building towards a complete "Crawl, Fix, & Validate" ecosystem for AEO that will launch fully in early February.

    Right now, the scanner is live if you want to check your own site's "AI readability."

    Happy to answer questions about the crawling methodology or the specific schema failures we saw in the wild.

    • JohnFenan hour ago
      > (often without the owner knowing)

      How can you tell this? Why do you call this the "accidental blockade"? Surely, at least some percentage of those sites are doing it intentionally.

      • aggeeinnan hour ago
        Fair question. We distinguish them based on the specificity of the rule. If a robots.txt file explicitly names GPTBot or CCBot, we count that as intentional. The accidental group consists of sites using generic User-agent: * disallows (often left over from staging) or legacy security plugins that block unknown user agents by default. We spot-checked a sample of these owners, and most were completely unaware that their 5-year-old config was actively blocking modern AI agents.
  • CableNinjaan hour ago
    Id be more curious on finding out what AI bots can access my site, so i could stop it.

    At the public disclosure of chatgpt i immediately went and added a block in my nginx config. I would ideally like to block them all.

    Im currently relying on UA and have a tiny if statement in my config that tells every ai ive blocked my server is simply a teapot

    • aggeeinnan hour ago
      The 418 status is a nice touch. We actually noticed that whack-a-mole issue across the entire dataset—keeping a static Nginx config synced with the explosion of new user-agents is proving difficult for most admins right now.

      If you're curious to stress-test the regex, feel free to drop the URL (or check my profile for email). I can run a quick pass with our crawler to see if it triggers the teapot response or if the headers manage to slip through.