← Back to context

Comment by dijit

5 days ago

Uncharitable.

Robots.txt is already the understood mechanism for getting robots to avoid scraping a website.

People often use specific user agents in there, which is hard if you don't know what the user agents are in advance!

  • That seems like a potentially very useful addition to the robots.txt "standard": Crawler categories.

    Wanting to disallow LLM training (or optionally only that of closed-weight models), but encouraging search indexing or even LLM retrieval in response to user queries, seems popular enough.

  • If you're using a specific user agent, then you're saying "I want this specific user agent to follow this rule, and not any others." Don't be surprised when a new bot does what you say! If you don't want any bots reading something, use a wildcard.

    • Yes, but given the lack of generic "robot types" (e.g. "allow algorithmic search crawlers, allow archival, deny LLM training crawlers"), neither opt-in nor opt-out seems like a particularly great option in an age where new crawlers are appearing rapidly (and often, such as here, are announced only after the fact).

    • Sure, but I still think it's OK to look at Apple with a raised eyebrow when they say "and our previously secret training data crawler obeys robots.txt so you can always opt out!"

      6 replies →