Managing Search Engines and AI Bots with robots.txt and llms.txt - GRAIsol Blog
Share this post
General

Managing Search Engines and AI Bots with robots.txt and llms.txt

1 min read

Search engine crawlers and AI bots hit every site, so we created clear guidelines on what they can access.

robots.txt Highlights

  • Allowed bots to crawl all public pages and static assets like /images/ and /portfolio/.
  • Blocked sensitive areas such as /api/, /_next/, and /admin/.
  • Added a sitemap reference and a modest Crawl-delay to reduce server strain.
  • Included specific rules for bots like GPTBot and ChatGPT-User to limit API and admin access.

llms.txt Guidelines

  • Outlined which site content AI models may quote, analyze, and index.
  • Prohibited training on client data, custom implementations, and pricing strategies.
  • Listed acceptable uses, such as answering questions about public services.
  • Provided contact emails for AI/LLM inquiries and permissions.

By publishing these files, we encourage good-faith indexing while protecting proprietary information and client privacy.

Share this post: