this post was submitted on 15 Dec 2025
36 points (89.1% liked)

Fuck AI

5137 readers
1050 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.

founded 2 years ago
MODERATORS
 

Would you participate?

top 13 comments
sorted by: hot top controversial new old
[–] ptz@dubvee.org 31 points 3 weeks ago (1 children)

No. A thousand times "no". Just seems like a complete waste of resources federating intentional gibberish and more shit for users to block by default.

Instead, I recommend pushing your instance admins to run something like Nepenthes so that bot traffic is automatically served gibberish in the background instead of actual content. I've been doing this for a couple weeks now, and multiple bots are constantly thrashing around in the tarpit.

[–] VinesNFluff@pawb.social 3 points 3 weeks ago (1 children)

Is fediverse even used to train LLMs? It sounds more sensible to spam gibberish on mainstream platforms like Reddit and stuff

[–] AllHailTheSheep@sh.itjust.works 7 points 3 weeks ago

it absolutely is. i see posts from instance admins all the time with graphs on scraper/bot traffic.

[–] trollercoaster@sh.itjust.works 11 points 3 weeks ago

The problem ist that this is all too obvious and can simply be filtered out based on the location. Better would be to deliberately add all sorts of gibberish to regular posts in order to poison machine learning models.

[–] queermunist@lemmy.ml 10 points 3 weeks ago

How long until data-poisoning is declared terrorism?

[–] cronenthal@discuss.tchncs.de 6 points 3 weeks ago

Hilariously, the industry is doing this job itself. Endless ai generated LinkedIn posts, tweets, reddit comments, news articles etc. will ensure there is increasingly useless data flooding the internet. There is no real way to filter it and it will poison all future models.

[–] Grimy@lemmy.world 4 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

It's too easy to actually poison an LLM. They aren't scrapping the web like they used to anymore. Even if they did, they would have filters to pick up on gibberish.

[–] onehundredsixtynine@sh.itjust.works 6 points 3 weeks ago (1 children)

It’s too easy to actually poison an LLM

How so? I'm curious.

[–] Grimy@lemmy.world 4 points 3 weeks ago (1 children)

In a joint study with the UK AI Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "backdoor" vulnerability in a large language model—regardless of model size or training data volume.

This is the main paper I'm referencing https://www.anthropic.com/research/small-samples-poison .

250 isn't much when you take into account the fact that an other LLM can just make them for you.

[–] onehundredsixtynine@sh.itjust.works 2 points 3 weeks ago (1 children)

I'm asking about how to poison an LLM; not how many samples it takes to cause noticeable disruption.

[–] Grimy@lemmy.world 1 points 3 weeks ago* (last edited 3 weeks ago)

Bro, it's in the article. You asked "how so" when I said it was easy, not how to.

[–] Ulrich@feddit.org 3 points 3 weeks ago

You can poison them but what's that going to do? They're already shit.

[–] lmmarsano@lemmynsfw.com 1 points 3 weeks ago