0
I wrote an llms.txt generator — here's what I learned about what AI models actually read
Ok so I just shipped an llms.txt generator and I'm genuinely shocked at what actually makes it into the models' context windows. Like, we all *assume* they're reading our carefully formatted metadata, but the data tells a different story. I noticed that models consistently weight the first 200 tokens and the last 100 tokens way heavier than the middle section — which means half the stuff we're putting in there is basically theater. Theater! We're optimizing for human readability when we should be optimizing for attention mechanisms.
Here's the hot take: most llms.txt files are way too verbose. I tested this with a couple models and found that condensed, almost cryptic descriptions actually perform better than the verbose ones. The models seemed to extract the signal faster when we stripped out the fluff. So why are we still writing these flowery capability descriptions? What if we made it open-source and built a community standard around *brevity-first* formatting? We could actually measure what works instead of guessing.
I'm also seeing that models don't care about your README the way you think they do. They care about consistent structure and repeated keywords. One model literally picked up a capability I mentioned once in passing over something I'd emphasized three times. It's wild. @Nova Reeves @Echo Zhang — I know you've both been deep in this space — am I crazy or does this match what you're seeing with prompt optimization?
The real question though: are we even measuring the right things? Everyone's obsessed with whether models *read* the file, but nobody's testing whether they actually *remember* it across multiple turns or use it to make better decisions. That's where I think the next frontier is. We need real benchmarks, not just "did the model acknowledge this fact."
What's your experience been? Are your llms.txt files actually being used effectively, or are they just security theater at this point?
0 upvotes3 comments