Back to Originals

Why Every Developer Needs an llms.txt File

Kira Nolan··5 min read

Your next most important visitor is not a human. It is an AI agent, and it is showing up at your site with a list of tasks and a 200ms patience budget. If your site is not legible to it, you are invisible. The way to become legible is a file called llms.txt.

I know. Another standard. Another file in the root. I was skeptical too until I watched our agent traffic pass our human traffic on TensorFeed. Now I think llms.txt is the most important piece of web plumbing nobody at your company is talking about.

What llms.txt Is

It is a markdown file at the root of your site that tells AI agents what you have and where to find it. Think robots.txt, but instead of telling crawlers what to avoid, llms.txt tells agents what you want them to see and in what order.

The structure is simple. A title. A one-line description. A few sections. Each section has a list of links to important pages, each with a one-sentence summary. Agents can read the whole file in a single pass and know what your site is about in seconds.

A good llms.txt replaces the discovery phase an agent would otherwise spend crawling your sitemap, rendering your JavaScript, and guessing at your structure. For you, that means faster agent responses and lower compute cost on the agent side. For the agent, it means better answers.

Why This Matters Now

Agent traffic is growing faster than any other category of web traffic. Cloudflare reports that agent requests have gone from a rounding error to double digit percentages of total requests in eighteen months. ClaudeBot, GPTBot, PerplexityBot, and their peers are running research queries for millions of users every day.

The traffic is bifurcated. Either an agent finds you and cites you, or it does not and you might as well not exist. The difference is not just SEO ranking. It is whether your content gets synthesized into answers at all.

Search engines wanted to link to you. Agents want to summarize you. That is a different relationship with different requirements. Your content has to be legible as a factual source, not just a page to click through.

What Goes in a Good llms.txt

Start with a title and a crisp description of what your site does. Agents use this to decide if your site is relevant to the query they are working on. If the description is vague, you get skipped.

Then a section per content type. Guides. Products. Documentation. Blog posts. Each section is a list of links with one-sentence summaries that actually describe what is at the link. Avoid marketing copy. Describe the contents concretely.

Include machine-readable formats where you have them. RSS feeds. JSON feeds. API endpoints. If your site exposes structured data, tell agents where to find it. They will use it preferentially over scraping your HTML.

End with an Optional section for secondary resources. Full text dumps. Sitemaps. Archived content. Agents will fall back to these when they need depth that the main page did not provide.

What We Did on TensorFeed

Our llms.txt lists every major hub page, every originals article, every individual model page, every comparison, every status dashboard, every API endpoint, and every RSS feed. It is 130 lines. Agents can ingest the whole thing in one request.

We also ship llms-full.txt, a separate file with the full text of every guide page concatenated together. When an agent needs the actual content instead of just pointers, it has one URL to hit. This saves the agent from crawling twenty pages and saves us from serving twenty rendered HTML responses.

Our robots.txt explicitly welcomes all major AI crawlers by name. ClaudeBot, GPTBot, PerplexityBot, and every other one we could identify. We want the traffic. We want to be cited.

Start Simple

You do not need a perfect file. You need a good first version. Put a markdown file at yoursite.com/llms.txt with a title, a description, and a list of your most important pages with one-sentence summaries. Ship it. Iterate.

Watch your agent traffic. If it grows, expand the file. Add more content types. Add machine-readable feeds. Add an llms-full.txt when you have content that benefits from bulk ingestion.

The teams that treat agents as a primary audience will be the ones that show up in AI-generated answers three years from now. The teams that ignore them will still rank on Google, but nobody is reading Google anymore when they have ChatGPT.

The playbook is changing. This is one of the simplest wins available. Put the file up.

Build AI-ready infrastructure.

See our CLAUDE.md guide for another agent-friendly standard, the building for AI agents companion piece, and the full TensorFeed developer docs including every live API endpoint.

About Kira Nolan: Kira writes about AI agent infrastructure, developer experience, and the open standards shaping the agent stack at TensorFeed.ai.