Skip to main content
Industry Updates

Which llms.txt Tool Actually Works? We Benchmarked 14

Jamie Watters 9 min read
Table of Contents

The llms.txt market has exploded. In the past year, more than a dozen tools have launched claiming to improve your AI search visibility — from free generators to enterprise platforms costing $489 per month. We tested 14 of them against a structured evaluation framework. The results reveal a market where most tools do one thing, a few do two things, and almost none complete the full loop that actually moves the needle.

The One-Thing Problem

Most AI visibility tools do a single thing: generate an llms.txt file. You enter your URL, the tool crawls your site, and it produces a file. For many businesses, this feels like a solved problem. You have a file. You deploy it. Done.

Except that file generation is step one of a four-step process. A generated file tells you nothing about whether it's any good, whether it actually improved your AI visibility, or what to do next. It's like getting a medical scan without a diagnosis — you have data, but no actionable information.

The tools that stop at generation leave you guessing. Did the file work? Is AI citing you more now? Are there underlying site issues that undermine the file entirely? Without answers to these questions, a generated llms.txt file is a hope, not a strategy.

The Full-Loop Framework

We evaluated all 14 tools against a four-step framework that represents the complete AI visibility workflow:

  1. Diagnose — Can the tool identify specific problems with your AI visibility? Not just "generate a file" but "tell you what's wrong and why."
  2. Optimise — Can the tool implement fixes or provide actionable, specific recommendations? Generic advice like "improve your content" doesn't count.
  3. Test — Can the tool verify that the changes actually worked? Can you re-scan and see measurable improvement?
  4. Repeat — Can the tool track progress over time? Does it show you a trajectory, not just a snapshot?

Most tools cover step one, partially. Some address step two. Almost none complete steps three and four. And without those final steps, you're flying blind.

Benchmark Results

We scored each tool across six dimensions: generation quality, validation capability, AI readability scoring, re-testing capability, JavaScript rendering, and improvement tracking. Maximum score: 30 points.

Tool Score (/30) Full Loop? Price
LLM.txt Mastery 18 Yes $4.95/mo
LLMrefs 15 Partial $29/mo
Otterly AI 14 Partial $49/mo
Peec AI 12 Partial $39/mo
Cairrot 11 Partial $29/mo
AIOSEO 4 No $49/mo

Table shows 6 of 14 tools tested. Remaining 8 scored between 3 and 10, with none completing the full loop.

The standout finding: the highest-scoring tool costs $4.95 per month. Several lower-scoring tools cost 5–10x more. Price does not correlate with capability in this market — a reflection of how new the category is and how differently vendors interpret the problem.

What Separates the Top Tools

JavaScript Rendering

Between 60% and 80% of modern websites use JavaScript frameworks. A tool that can't render JavaScript literally cannot see the content on most websites. It's generating an llms.txt file based on an empty page.

Of the 14 tools tested, only 3 offered full JavaScript rendering. The rest fetched raw HTML, missing all dynamically rendered content. For any site built with React, Vue, Angular, or a similar framework, this is a disqualifying limitation.

Readability Scoring Specificity

Several tools offer an "AI friendliness" score — a single number that claims to represent how well your site works for AI. These scores are largely meaningless without specific, factor-level breakdowns. A score of 72 out of 100 tells you nothing about what to fix.

The more useful tools evaluate specific factors — structured data presence, content depth, authorship signals, entity recognition, freshness indicators — and tell you exactly which ones are weak. The MASTERY-AI Framework evaluates 27 specific factors across 8 pillars, each with actionable recommendations.

Re-Testing Capability

You make changes. You implement the recommendations. Now what? Without re-testing, you have no way to know whether your changes worked. You can't see the delta between your previous score and your current one.

Only 4 of the 14 tools offered any form of before-and-after comparison. The rest treated each scan as an isolated event with no connection to previous results.

The Free Generator Trap

Free llms.txt generators are everywhere. They're the entry point for most businesses exploring AI visibility. And they come with predictable limitations:

  • No JavaScript rendering — Free tools don't run headless browsers. If your site uses a JavaScript framework, the generated file will be incomplete or empty.
  • No quality scoring — Every page gets listed equally. Your flagship product page carries the same weight as your cookie policy.
  • Bloated output — Without filtering, free generators produce files that list every URL they find, creating noise that can degrade AI parsing performance.
  • Lead magnets — Many free generators exist primarily to capture email addresses. The tool is the marketing, not the product.

A free generator can tell you whether you have an llms.txt file. It can't tell you whether the file is any good, whether your site is actually ready for AI, or what to do next. For businesses that are serious about AI visibility, free generation is the starting point — not the solution.

How to Choose the Right Tool

Based on our benchmark data, here's what to evaluate when selecting an AI visibility tool:

  1. Does it diagnose? — Can the tool tell you specifically what's wrong with your AI visibility, not just generate a file?
  2. Does it render JavaScript? — If your site uses any modern framework, this is non-negotiable.
  3. Does it score specifically? — Generic "AI friendliness" scores are unhelpful. Look for factor-level breakdowns with actionable recommendations.
  4. Can you re-test? — After implementing changes, can you re-scan and measure improvement?
  5. Does it track progress? — Can you see your trajectory over time, not just isolated snapshots?
  6. Is the price sustainable? — Enterprise pricing ($49–$489/month) may be justified for large teams. For solopreneurs and small businesses, the cost needs to reflect the value delivered.

The AI visibility tool market is still maturing. Prices will stabilise, features will consolidate, and the tools that deliver the full loop — diagnose, optimise, test, repeat — will separate from the ones that just generate files. Choose accordingly.

Most tools stop at file generation. That's step one of four.

AImpactScanner diagnoses, scores, and tracks your AI visibility across 27 factors.