LLMs.txt has generated lots of heat—and little light. Here’s a pragmatic, evidence‑based take: what it is, what it isn’t, and how to use it (if at all) without burning dev hours.
TL;DR
- Major AI search experiences do not rely on LLMs.txt today. Focus on answer‑centric content, clean structure, and technical hygiene.
- It can be useful as a curated “starting map” for agent and RAG workflows.
- If you publish one, keep it minimal, accurate, non‑sensitive, and aligned with your sitemap. Treat it as optional housekeeping.
What LLMs.txt tries to do
LLMs.txt is a plain‑text file at your web root that lists high‑value pages (e.g., docs, FAQs, benchmarks) and offers light hints (update cadence, canonical locations). It is closer to a curated sitemap than a robots.txt directive—informational, not enforceable.
Current state: crawled, mostly ignored by AI search
- Public guidance for AI Overviews: follow normal SEO best practices; LLMs.txt isn’t used for inclusion or ranking.
- Large‑scale crawler logs show occasional fetches of
/llms.txt
but no meaningful effect on inclusion, ranking, or citation in AI answers. - Industry commentary broadly aligns: tiny effort, tiny impact—don’t expect visibility gains.
Where LLMs.txt can help (today)
- Agent‑first browsing: Give autonomous/semiautonomous agents a curated jump‑off list for definitive resources (docs, changelogs, pricing, comparisons).
- RAG seed lists: Provide initial URLs for chunking/indexing in retrieval pipelines.
- Enterprise assistants: Standardize “where to start” across microsites and subdomains.
Risks and limitations
- Voluntary and unenforceable: Providers aren’t obligated to honor it.
- Redundant: Good pages are already discoverable via HTML, sitemaps, and links.
- Misconfiguration risk: Don’t surface sensitive, ephemeral, or low‑quality URLs.
Minimal LLMs.txt (quick start)
version: 1
urls:
- https://example.com/docs/
- https://example.com/pricing/
- https://example.com/blog/
notes: canonical resources; updated monthly
Guidelines:
- Keep it short and consistent with your public sitemap.
- Only list pages you would confidently recommend to an agent or researcher.
- Add a brief notes line for cadence or canonicals; avoid proprietary data.
Adoption checklist (do no harm)
- Confirm there’s no expectation of SEO/AI visibility gains from LLMs.txt alone.
- Mirror existing canonical pages; avoid pre‑release/beta/private endpoints.
- Review quarterly (or when IA changes) to prevent drift.
- Document policy internally (what qualifies, who updates, when).
What actually moves the needle for GEO
- Answer‑centric content: Clear sections that directly answer prompts, with evidence and quotable statements.
- Comparison‑first pages: Transparent criteria and extractable tables for “X vs Y” use cases.
- Structured signals: Clean titles, headings, summaries, tables, and appropriate schema.
- Technical foundation: Fast, crawlable, canonical pages with strong internal linking.
FAQ
Should we add LLMs.txt?
Maybe. If it’s low‑effort and you maintain it, it can help agent/RAG workflows. Don’t expect AI search uplift.
Is it like robots.txt?
No. Robots.txt provides crawl directives; LLMs.txt is optional guidance. It doesn’t control access.
Does it help AI Overviews?
No current evidence. Focus on content quality, structure, and standard technical SEO.
How do we measure impact?
Track agent/RAG usage internally. For AI search, monitor mentions/citations and traffic patterns independent of LLMs.txt.
Key takeaways
- Treat LLMs.txt as optional, agent‑oriented metadata—not a GEO lever.
- If you publish it, keep it minimal, accurate, and aligned with your sitemap.
- Invest most effort in answer‑centric content, structured comparisons, and technical excellence.
Last updated: 2025‑10‑09
About the Author
Vladan Ilic
Founder and CEO