this post was submitted on 02 Jan 2026
266 points (97.8% liked)

Fuck AI

5137 readers
1072 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.

founded 2 years ago
MODERATORS
 

For days, xAI has remained silent after its chatbot Grok admitted to generating sexualized AI images of minors, which could be categorized as violative child sexual abuse materials (CSAM) in the US.

you are viewing a single comment's thread
view the rest of the comments
[–] DaTingGoBrrr@lemmy.world 6 points 6 days ago (1 children)

Is no one going to question HOW and WHY Grok knows how to generate CSAM?

This is fucking disgusting and both the user and X should be held accountable.

[–] IchNichtenLichten@lemmy.wtf 8 points 6 days ago (1 children)

I agree that it’s disgusting. To answer your question, it doesn’t know anything. It’s assigning probabilities based on its training data in order to create a response to a user prompt.

[–] DaTingGoBrrr@lemmy.world 3 points 6 days ago* (last edited 6 days ago) (3 children)

Yes I know that it doesn't "know" anything, but where did it get that training data and why does it have CSAM training data? Or does it just generate regular porn and add a kid face?

[–] IchNichtenLichten@lemmy.wtf 6 points 6 days ago

These companies tend to not say how they train their models, partially because much of it is stolen but the data is of pretty much everything. The LLM will generate a response to any prompt so if it can be used to put a celebrity in lingerie, it can also be used to do the same with a child. Of course there are guardrails, but they’re weak and I hope X gets sued into oblivion.

[–] HK65@sopuli.xyz 3 points 5 days ago

There are two answers to that, both equally valid.

One is that it extrapolates based on knowing what a naked adult looks like compared to a clothed adult, and how a child looks like compared to an adult, it can "add those vectors" and figure out how a naked child looks like.

The other is that one of the biggest porn datasets that most of these will have in their training data has recently been taken down because it had a bunch of CSAM in it. Ironically, how it happened was that an independent guy uploaded it to Google Cloud, and Google flagged and banned the guy for it.

The dataset would not have been taken down if it wasn't for the guy doing the rounds afterwards though. Google didn't care beyond banning a user.