I set my server to limit requests per hour from the same IP to slow them
down, and I have code to detect bots and redirect their sessions to a low
impact catch page.  It’s not that hard to control, but lately I have
noticed old tricks no longer work as well.  AI arms race.  But I always
believed publishing publicly would eventually cause the content to enter
the public domain.
On Tue, Sep 16, 2025 at 9:02 PM Cameron Kaiser via cctalk <
cctalk(a)classiccmp.org> wrote:
  For those of you who run vintage
computing-related info sites, have you
 noticed all of the LLM scraper activity?    AI services are using the LLM
 scrapers to populate their knowledge bases.
 
 A massive, massive IP filter. There has been some collateral damage, but
 unfortunately I don't think this is avoidable. They're a plague.
 --
 ------------------------------------ personal:
 
http://www.cameronkaiser.com/ --
   Cameron Kaiser * Floodgap Systems * 
www.floodgap.com *
 ckaiser(a)floodgap.com
 -- Time is an illusion. Lunch time, doubly so. -- Douglas Adams
 ---------------