this post was submitted on 14 May 2025
92 points (100.0% liked)

TechTakes

1859 readers
264 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 1 day ago* (last edited 1 day ago) (1 children)

In my experience with bots, a portion of them obey robots.txt, but it's tricky to find the user agent string that some bots react to.

So I recommend having a robots.txt that not only target specific bots, but also tell all bots to avoid specific paths/queries.

Example for dokuwiki

User-agent: *
Noindex: /lib/
Disallow: /_export/
Disallow: /user/
Disallow: /*?do=
Disallow: /*&do=
Disallow: /*?rev=
Disallow: /*&rev=
[–] [email protected] 3 points 23 hours ago (1 children)

Would it be possible to detect the gptbot (or similar) of their user agent, and server them different data?

Can they detect that?

[–] [email protected] 7 points 22 hours ago* (last edited 22 hours ago) (2 children)

yes, you can match on user agent, and then conditionally serve them other stuff (most webservers are fine with this). nepenthes and iocaine are the current preferred/recommended servers to serve them bot mazes

the thing is that the crawlers will also lie (openai definitely doesn't publish all its own source IPs, I've verified this myself), and will attempt a number of workarounds (like using residential proxies too)

[–] [email protected] 2 points 10 hours ago* (last edited 10 hours ago)

Generating plausible-looking gibberish require resources. Giving any kind of response to these bots is a waste of resources, even if it's giberish.

My current approach is to have a robots.txt for bots than honor it. And drop all traffic during 24h for IPs used by bots that ignore robots.txt or misbehave.

[–] [email protected] 3 points 22 hours ago

Can they detect that they're being served different content though?