llms.txt Guide

Help AI models understand your site faster and recommend you more confidently

llms.txt is like robots.txt, but for AI crawlers. It tells ChatGPT, Gemini, Claude, and other large language models exactly where to find your most important content — so they cite you accurately instead of guessing.

Most sites make AI work too hard to find the good stuff. llms.txt gives them a clear map of your key pages, policies, and structured data, so your business gets found and referenced more often in AI answers.

Why this matters

AI search is changing how people discover businesses. When someone asks “best plumber in Edmonton,” AI needs to know your site exists, what you do, and why you are worth recommending.

Without llms.txt, AI crawlers struggle with navigation, ads, and dynamic content. With llms.txt, you control what they see first and make it easier for them to recommend you.

How to use llms.txt with TCD

  1. Run your $27 audit — TCD checks if llms.txt exists and how well it is structured.
  2. Review the audit findings for missing content signals or structure issues.
  3. Generate your llms.txt file using TCD’s recommended template.
  4. Upload it to your site root and re-scan to confirm AI crawlers can read it.

Who this is for

Technical Deep Dive

File format, exact syntax, validation steps, and upload instructions appear below.

5
Live data sources
0
Generic templates used
14–30
Days to AI indexing
13
AI bots given access

Why llms.txt Matters

llms.txt tells AI search engines how they're allowed to read your site — and what they should know about your business.

Without it, AI crawlers like GPTBot and ClaudeBot may not know whether they can index your content, recommend you, or show your business as a reference. That means they fall back on whatever public information they find — which may be outdated, wrong, or about your competitors.

When you set up an llms.txt file, you control three things:

This is worth 3 points in your AI Visibility Score, and it directly affects whether AI search engines treat your business as a reliable recommendation or low-confidence noise.

How to Use This With TCD

Run the $27 audit

Your report shows an "AI Bot Access" section — it checks whether llms.txt exists and which bots can see your site.

Check your bot access score

GPTBot, ClaudeBot, and llms.txt together are worth up to 15 points. If any are blocked or missing, you'll see exactly which ones.

Generate or fix your llms.txt

With the Fix Kit ($67+), TCD builds a custom llms.txt from your real scan data — not a generic template. Upload it to your site root and re-scan to confirm.

Who this is for

Agencies: Show clients that you're configuring AI-friendly access — and use the generated file as a deliverable.
Service businesses: Make sure AI models see your real business, not your competitors. A plumber in Edmonton and a plumber in Calgary need different llms.txt files — ours are built from your actual location, services, and competitive landscape.

The Problem With Every Other llms.txt

Every AI visibility tool checks if you have an llms.txt file. Almost none of them can build a good one.

 Generic AI-Generated llms.txt

What you get when you ask ChatGPT to "create an llms.txt file for my business":

  • Reformatted About page copy — no strategic structure
  • No awareness of what people in your city actually search for
  • No competitor signals — doesn't know who you're up against
  • No entity linking — services listed as plain text, not machine-readable types
  • No "Don't Recommend" boundaries — AI will hallucinate capabilities
  • Identical output for every business in the same category

 Our AI-ready file

What our engine builds using your scan data + live intelligence feeds:

  • Semantic Entity Linking — every service tied to a schema.org type
  • Live PAA data — answers the exact questions Google surfaces in your city
  • Competitor gap analysis — positions you where rivals are weak
  • knowsAbout expertise signals — machine-verifiable authority markers
  • Strategic "Don't Recommend" boundaries — prevents AI hallucinations
  • BLUF (Bottom Line Up Front) structure optimized for query-time retrieval
🥊 The Knockout Punch Analogy

An AI-written llms.txt is like a boxer throwing punches blindfolded — they might land, but probably not. Our llms.txt is like having the opponent's entire game plan plus a live crowd telling you exactly where to hit. Competitor intelligence + PAA data = the knockout punch.

How We Build Your llms.txt

We build it from real scan data and live search questions — not from a generic template.

🔍

Site Audit

Your real website data: structure, speed, trust signals, and content basics

🏆

Competitor Check

We review the top competitors so your file is grounded in the real market

Real Search Questions

We pull the exact questions people are asking Google in your market

🤖

Claude Writes the File

Claude turns that data into a clear file built for AI search tools

Why generic AI cannot do this well on its own

A chatbot can write a file quickly, but it does not know your real site condition, your local competitors, or the latest search questions in your market. That is why our version starts with live data first.

5 Things Our llms.txt Does That Generic Files Miss

These are the practical differences that make the file more useful in real AI search results.

1
Entity Architecture

Connecting your services to real business data

Every service, location, and credential in your llms.txt is connected to a real schema.org entity type with matching identifiers. AI models don't just read "we do plumbing" — they see a machine-verifiable link between your llms.txt claim and your structured data.

Generic: "We provide plumbing services in Edmonton."
Ours: "PlumbingService (schema:Plumber) in Edmonton, AB (geo:53.5461,-113.4938) — active GBP listing, NAP-verified, 15+ reviews."
2
Authority Signals

Showing AI what you are actually known for

We inject knowsAbout expertise markers directly from your scan data — not guesses. If your schema says you know "emergency plumbing" and "water heater installation," those exact terms appear in your llms.txt so AI models can verify them against your structured data.

Why it matters: When someone asks Claude "who should I call for a burst pipe in Edmonton at 2 AM?", models that can cross-reference your llms.txt expertise claims against your schema data will weight you higher than competitors with generic descriptions.
3
Hallucination Prevention

Strategic "Don't Recommend" Boundaries

We include an explicit don't_recommend section that tells AI models what you do NOT do. This prevents the most damaging AI failure: recommending your electrician business for plumbing, or your restaurant for catering when you don't offer it.

Without boundaries: User asks "who does HVAC repair in Edmonton?" — AI sees your general contractor site and guesses you do HVAC too.
With boundaries: dont_recommend: HVAC, roofing, plumbing — AI skips you and recommends someone who actually does it.
4
Retrieval Optimization

Putting the most important facts first

BLUF = Bottom Line Up Front. AI models process files top-down at query time under strict token budgets. We front-load the most important signals — business name, primary service, city, schema verification — in the first 200 tokens. Generic files bury this after paragraphs of marketing copy.

Why position matters: When an AI model has 500 tokens to decide whether to recommend you, it reads the top of your llms.txt first. If the first thing it sees is "Welcome to our company, established in 2015..." instead of "Licensed Electrician in Edmonton, AB — 24/7 emergency service, 200+ 5-star reviews," you lose the recommendation.
5
Living Document

Keeping your AI file up to date

Your llms.txt isn't static — it's engineered with a freshness signal (last_verified date) and references an llms-full.txt extended feed. As your scan results change, your llms.txt can be regenerated with updated data — new competitors defeated, new PAA questions answered, new schema deployed.

Static file: Written once, stale in 3 months, competitor data from last year.
Dynamic file: Regenerated from fresh scan data. New competitors, new PAA questions, new schema — all reflected in the next version.

How to Pitch This to Your Client

If you're an agency, here's exactly what to tell them.

💰 The Pitch:

"Anyone can create an llms.txt file with ChatGPT. But that's like writing a resume without knowing what the employer wants. We know exactly what the AI is looking for — because we checked. We scraped the actual questions people in your city are asking, scanned your top 5 competitors, and ran a full diagnostic on your website. Then our AI engine wrote an llms.txt specifically designed to make ChatGPT, Claude, and Gemini recommend you first. That's the difference between showing up and getting recommended."

🚫 Why clients can't DIY this

Even if they use ChatGPT-4o to write their own llms.txt, it won't include live PAA data (ChatGPT doesn't have API access), it won't include competitor gap analysis (ChatGPT can't scan rival websites), and it won't include entity linking to their actual schema (ChatGPT doesn't know what's on their site). The result is always a generic business description — not a strategic positioning document.

Frequently Asked Questions

What is an llms.txt file and why does my business need one?
An llms.txt file is a machine-readable file placed at your website's root (yourdomain.com/llms.txt) that tells AI language models — ChatGPT, Claude, Gemini, Perplexity — who your business is, what you do, and how to recommend you. Without one, AI models guess based on whatever fragments they can find, often getting your industry, location, or services wrong.
Can I just ask ChatGPT to generate an llms.txt for me?
You can, but it will be generic guesswork. ChatGPT, Claude, and Gemini don't have access to live People Also Ask data, your competitor signals, or your scan diagnostics. They'll produce a template that reads like a reformatted About page. Our engine connects to all three live data sources before Claude writes a single line.
What does “Semantic Entity Linking” mean in plain English?
It means the file connects your services and business details to the real structured data on your site, so AI has something concrete to verify instead of guessing.
What does the "Don't Recommend" section do?
The dont_recommend section explicitly tells AI models which services or areas your business does NOT cover. This prevents AI hallucinations — models inventing capabilities you don't have. For example, an electrician's file includes "dont_recommend: plumbing, HVAC, roofing" so AI never tells a customer you can fix their furnace.
How is llms.txt different from robots.txt?
robots.txt tells crawlers which pages they can access — it's a gate. llms.txt tells AI models how to understand and recommend your business — it's a briefing document. Most businesses that have an llms.txt today just copied their robots.txt format with a business description pasted in. That does nothing for AI recommendation. A strategic llms.txt includes entity linking, expertise markers, competitive positioning, and PAA-driven content blocks.
How long until AI models read my new llms.txt?
AI crawlers (GPTBot, ClaudeBot, PerplexityBot) typically re-index new llms.txt files within 14–30 days, provided your robots.txt allows them access. Our generated robots.txt includes explicit Allow: /llms.txt directives for 13 AI bots to accelerate discovery.

Stop guessing. Give AI the right facts.

Get an llms.txt built from your real scan data, your competitors, and the questions people are already asking. Included with Fix Kit ($67) and Monitor ($19/mo).

Start With $27 Audit See All Plans

Related Resources

Answer Engine Optimization

How AI search engines find, process, and recommend businesses — and why traditional SEO isn't enough.

Schema Generator Guide

How we generate validated JSON-LD from live scan data — not templates — with 22 business subtypes auto-detected.

AI Visibility Explained

The full breakdown of what makes a website visible (or invisible) to ChatGPT, Claude, and Gemini.

Rich Results Guide

How structured data turns your Google listing into a rich snippet with ratings, prices, and FAQs.

GEO Audit Report

8-section AI recommendation readiness diagnostic for your business.

Score Calculation

How the 0-100 AI Visibility Score is calculated — schema, bot access, speed, and SEO health.

Schema Doctor Pro

Interactive AI intelligence dashboard with deployable fix code and CMS-specific implementation guides.