this post was submitted on 25 Feb 2026
432 points (99.3% liked)

Funny

14018 readers
602 users here now

General rules:

Exceptions may be made at the discretion of the mods.

founded 2 years ago
MODERATORS
top 7 comments
sorted by: hot top controversial new old
[–] Tar_alcaran@sh.itjust.works 45 points 5 days ago* (last edited 5 days ago) (1 children)

Also pictured here: Anthropic stating out loud their models will just give out all the "secret" and "secured" internal data to anyone who asks.

Of course, that's by design. LLMs can't have any barrier between data and instructions, so they can never be secure.

[–] Hackworth@piefed.ca 17 points 5 days ago

Distillation is using one model to train another. It's not really about leaking data.

Claude was used to generate censorship-safe alternatives to politically sensitive queries like questions about dissidents, party leaders, or authoritarianism, likely in order to train DeepSeek’s own models to steer conversations away from censored topics

But you're right, prompt injection/jailbreaking is still trivial too.

[–] Mika@piefed.ca 35 points 5 days ago

The AI companies are inbreeding intentionally now? Wonderful!

[–] 0_o7@lemmy.dbzer0.com 8 points 5 days ago* (last edited 3 days ago)

When they steal: Innovative approach to knowledge acquisition

When others steal: A threat to free market by IP violation

[–] Hackworth@piefed.ca 10 points 5 days ago* (last edited 5 days ago)
[–] mindbleach@sh.itjust.works 8 points 5 days ago

In undue fairness, there is a difference between turning text files into a chatbot, and exfiltrating that chatbot. One is transformative, and the other is making a megaphone out of some string, a squirrel, and a megaphone.

But if I don't give a shit about companies doing math on Disney DVDs I'm not about to give a shit about them hoarding their big pile of numbers. I'm jazzed when source code leaks for things written by people.