Llmtext – an open source toolkit for llms.txt adoption

3 days ago 2

Reduce hallucinations in Cursor, Claude, and other LLMs

CREATE llms.txt FROM SOURCES

Generate an llms.txt with accompanied markdown content from multiple sources, to help agents access everything easily.

VIEW ON GITHUB

VALIDATE llms.txt FILES

Check if your llms.txt file follows the official specification from llmstxt.org

TURN ANY llms.txt URL INTO A DEDICATED MCP SERVER

POPULAR llms.txt MCP SERVERS

LOADING...

RANK SOURCE TOKENS VOLUME USERS INSTALL

FREQUENTLY ASKED QUESTIONS

GETTING STARTED

What is llms.txt?

llms.txt is a standardized file format (similar to robots.txt) that helps AI agents and LLMs efficiently navigate and understand your website's content. It provides a structured table of contents with links to markdown or plain text versions of your pages.

Why is llms.txt important?

As AI traffic grows exponentially, traditional web scraping becomes inefficient and costly. llms.txt allows you to serve AI-friendly content directly, reducing latency and bandwidth while improving the quality of information AI agents can retrieve from your site.

What does the llms.txt MCP allow me to do?

The llms.txt MCP (Model Context Protocol) server turns any valid llms.txt file into a structured resource that AI assistants like Claude can directly access. This means developers can reference documentation, APIs, or content from any llms.txt-enabled website without hallucination or manual copying.

What's the difference between llms.txt MCP and Context7?

While Context7 uses vector search to find relevant content, llms.txt MCP provides structured navigation through a table of contents. This allows for more precise retrieval and reasoning over content organization, often more efficiently than embedding-based search for well-documented resources.

Who should use "Create"?

The Create tool is perfect for website owners whose CMS doesn't support markdown output. It generates an llms.txt file AND converts all your pages to markdown automatically using the Parallel Extract API, saving you from manual conversion work.

TECHNICAL QUESTIONS

What format should my llms.txt file follow?

Your llms.txt must follow the official specification. Key requirements include:

  • Served at your root domain: yoursite.com/llms.txt
  • Content-Type must be text/plain or text/markdown
  • Contains markdown-formatted links to plain text/markdown pages
  • Keep total size under 10k tokens
  • Each linked page should also be under 10k tokens
Why is my llms.txt failing validation?

Common validation failures include:

  • Wrong location: File must be at root (yoursite.com/llms.txt)
  • Wrong content-type: Must respond with text/plain or text/markdown
  • Links return HTML: Linked pages must return plain text/markdown
  • Too large: File exceeds 10k tokens
  • Full docs in one file: Should link to separate pages

Use our CHECK tool to identify specific issues.

How do I fix "links return HTML" errors?

Your llms.txt should link to markdown or plain text versions of pages, not HTML. Options:

  1. Configure your CMS to serve markdown versions at alternate URLs
  2. Use our CREATE tool to automatically convert and host markdown versions
  3. Set up content negotiation to return markdown when Accept header is text/markdown
What's the recommended token limit for llms.txt files?

Keep your llms.txt file under 10k tokens. The file itself should be compact—a table of contents, not the full content. Each page it links to should also stay under 10k tokens. This ensures efficient context usage and faster processing by AI agents.

Do I need to host markdown versions of my pages?

Yes, the URLs in your llms.txt must return content with Content-Type text/plain or text/markdown. If your site only serves HTML, use our CREATE tool which automatically scrapes and converts your pages to markdown using the Parallel Extract API.

PRACTICAL USAGE

How do I install an MCP server?

Click any "⬇ INSTALL" button on our site to visit installthismcp.com, which provides step-by-step instructions for installing MCP servers into Claude Desktop or other compatible AI assistants.

What makes a "high quality" llms.txt?

A high-quality llms.txt has:

  • Clear structure: Logical organization with descriptive titles
  • Concise entries: Brief but informative descriptions for each link
  • Proper scope: Links to pages, not entire docs in one file
  • Accessible content: All linked pages return markdown/plain text
  • Right size: Both llms.txt and linked pages under 10k tokens
  • Updated regularly: Reflects current site structure
How often should I update my llms.txt?

Update your llms.txt whenever you add, remove, or significantly reorganize content. For documentation sites, this might be with each release. For blogs, consider updating monthly or when adding major content sections.

What happens if my llms.txt changes after creating an MCP?

MCP servers fetch content dynamically, so changes to your llms.txt are reflected immediately. There's no need to "reinstall" or update the MCP—it always reads the current version of your llms.txt file.

Can I have multiple llms.txt files for different sections?

The standard llms.txt lives at your domain root, but you can create llms.txt files for subdomains (e.g., docs.yoursite.com/llms.txt). Each subdomain's llms.txt can have its own MCP server.

COMPARISON & INTEGRATION

How is this different from RAG or vector search?

Vector search (RAG) finds semantically similar content but loses document structure. llms.txt provides a navigable table of contents, allowing AI to reason about organization and relationships between topics. Both approaches are valuable—llms.txt excels for well-structured documentation, while RAG works better for unstructured knowledge bases.

Should I use llms.txt or a sitemap?

Use both! Sitemaps help search engines discover URLs. llms.txt provides AI-optimized content structure with markdown/text links. They serve complementary purposes—sitemaps for discovery, llms.txt for efficient AI consumption.

Which AI tools support llms.txt?

Any AI assistant with MCP support (like Claude Desktop) can use llms.txt via our MCP servers. Additionally, AI agents and browsers that follow the llmstxt.org standard can directly read these files. The ecosystem is growing rapidly.

Can I use this with my existing CMS?

Yes! If your CMS supports markdown export or custom content types, you can generate llms.txt directly. If not, use our CREATE tool to automatically generate both the llms.txt and markdown versions from your existing HTML content.

BUSINESS & STRATEGY

Why should my company invest in llms.txt?

As AI traffic grows, companies with llms.txt-enabled sites will:

  • Reduce server load from inefficient scraping
  • Improve AI-generated recommendations about their products
  • Enable developers to integrate with their APIs/docs without errors
  • Prepare for the AI-first internet where agents are primary users

Early adoption positions your content for the next wave of web traffic.

What are the SEO implications of llms.txt?

llms.txt is complementary to traditional SEO. It doesn't affect traditional search rankings but optimizes your site for "agent search optimization" (ASO)—making your content more discoverable and accurately represented by AI agents and chatbots, which are becoming major traffic sources.

Is llms.txt a web standard or just a proposal?

llms.txt is an emerging community standard defined at llmstxt.org. While not yet an official W3C standard, it's being rapidly adopted by major documentation sites and AI tool developers. Think of it like robots.txt in its early days.

Who else is using llms.txt?

Check our 03 INSTALL tab to see popular sites already using llms.txt. The list includes major documentation platforms, API providers, and developer tools. You can browse by popularity and tokens ingested.

Read Entire Article