this post was submitted on 02 Jan 2026
266 points (97.8% liked)
Fuck AI
5137 readers
1072 users here now
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Yes I know that it doesn't "know" anything, but where did it get that training data and why does it have CSAM training data? Or does it just generate regular porn and add a kid face?
These companies tend to not say how they train their models, partially because much of it is stolen but the data is of pretty much everything. The LLM will generate a response to any prompt so if it can be used to put a celebrity in lingerie, it can also be used to do the same with a child. Of course there are guardrails, but they’re weak and I hope X gets sued into oblivion.
There are two answers to that, both equally valid.
One is that it extrapolates based on knowing what a naked adult looks like compared to a clothed adult, and how a child looks like compared to an adult, it can "add those vectors" and figure out how a naked child looks like.
The other is that one of the biggest porn datasets that most of these will have in their training data has recently been taken down because it had a bunch of CSAM in it. Ironically, how it happened was that an independent guy uploaded it to Google Cloud, and Google flagged and banned the guy for it.
The dataset would not have been taken down if it wasn't for the guy doing the rounds afterwards though. Google didn't care beyond banning a user.
my concern as well https://medium.com/@russoatlarge_93541/canadian-child-protection-group-uncovers-abusive-content-in-academic-ai-dataset-9cc13f88701e
or maybe more relevant:
https://www.404media.co/laion-datasets-removed-stanford-csam-child-abuse/