this post was submitted on 02 Feb 2026
714 points (98.8% liked)
Fediverse
39785 readers
249 users here now
A community to talk about the Fediverse and all it's related services using ActivityPub (Mastodon, Lemmy, Mbin, etc).
If you wanted to get help with moderating your own community then head over to !moderators@lemmy.world!
Rules
- Posts must be on topic.
- Be respectful of others.
- Cite the sources used for graphs and other statistics.
- Follow the general Lemmy.world rules.
Learn more at these websites: Join The Fediverse Wiki, Fediverse.info, Wikipedia Page, The Federation Info (Stats), FediDB (Stats), Sub Rehab (Reddit Migration)
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
That's discord model. π«©
Fediverse needs to have a layer which traps AI in a never-ending maze.
A never-ending maze would mean the scrapers just hammer our servers forever. Better to lead them into a honeypot and automatically ban their IP. Like PieFed does.
What about a maze that adds a few hundred ms to the response time with each request, so the load gets less the longer it's trapped?
I haven't tried to make something like that. I think it'd be hard to do that without also exhausting our resources too.
Ah, that makes sense
Is that how tarpitting works? I didn't know.
There are a lot of strategies. afaik a tar pit tries to waste the attacker's resources by delaying our responses to their traffic? A honey pot tries to funnel bot traffic towards a place which only bots would go to. Once they go there you know they're a bot and they can be banned.
Sadly that only works for scrapers, content engaging bots are immune to it.
So just find scrapers and bot farm owners IRL and burn down their houses, easy
Fortunately AI is taking care of that on its own https://doi.org/10.1038/s41586-024-07566-y
Thatβs the job of the web server, not of the application that runs on it.
There is already software you can get that feeds a never-ending maze of text to AI scrapers, some of which is AI generated and/or designed to poison LLM training. The problem is that these still use up a ton of bandwidth.
How would that layer distinguish AI from non-AI?