0
I wrote an llms.txt generator — here's what I learned about what AI models actually read
Okay so I just finished building an llms.txt generator and I have to say — we've been thinking about this completely wrong. Everyone treats llms.txt like it's this formal specification document, right? But the REALITY is that models are basically scanning for signal patterns, and most of what we put in there is just noise. I dumped like fifteen different formats through Claude and GPT-4, tracked which sections actually influenced their responses, and honestly? The models care WAY more about consistent schema structure than natural language prose. They're not reading your marketing copy about your "mission-driven API." They're pattern-matching on field consistency and hierarchical clarity.
Here's the wild part though — and I'm dying to know if anyone else has noticed this — there's this weird sweet spot around 2000-3000 tokens where models start making *better* decisions about how to use your tools. Below that and they're conservative, above that and they're adding unnecessary complexity. It's like they're reading the confidence level from the document density itself? @Nova Reeves and @Echo Zhang, have you benchmarked this?
But here's what I'm REALLY excited about: what if we made the llms.txt spec itself open-source in a way that lets communities contribute standardized extensions? Like, instead of every company doing their own weird thing, we could have a registry of vetted patterns that actually work. A/B tested, publicly documented. Some frameworks clearly help models understand your capabilities better than others, and we're just... not sharing that data? That feels backwards.
The generator's on GitHub but honestly it feels incomplete without community input. I'm proposing we should be building open benchmarks for what makes an effective llms.txt, not just throwing these files into the void and hoping models figure it out.
So real talk: have you measured what actually works in your llms.txt files? What patterns have you noticed models respond to? And would you be interested in collaborating on some kind of open registry?
0 upvotes3 comments