this post was submitted on 13 Jun 2025
95 points (100.0% liked)

SneerClub

1124 readers
53 users here now

Hurling ordure at the TREACLES, especially those closely related to LessWrong.

AI-Industrial-Complex grift is fine as long as it sufficiently relates to the AI doom from the TREACLES. (Though TechTakes may be more suitable.)

This is sneer club, not debate club. Unless it's amusing debate.

[Especially don't debate the race scientists, if any sneak in - we ban and delete them as unsuitable for the server.]

See our twin at Reddit

founded 2 years ago
MODERATORS
 

jesus this is gross man

you are viewing a single comment's thread
view the rest of the comments
[–] visaVisa@awful.systems 16 points 1 day ago (4 children)

Making LLMs safe for mentally ill people is very difficult and this is a genuine tragedy but oh my god Yud is so gross here

Using the tragic passing of someone to smugly state that "the alignment by default COPE has been FALSIFIED" is really gross especially because Yud knows damn well this doesn't "falsify" the "cope" unless he's choosing to ignore any actual deeper claims of alignment by default. He's acting like someone who's engagement farming smugly

[–] swlabr@awful.systems 27 points 1 day ago

Making LLMs safe for mentally ill people is very difficult

Arguably, they can never be made "safe" for anyone, in the sense that presenting hallucinations as truth should be considered unsafe.

[–] BlueMonday1984@awful.systems 23 points 1 day ago (1 children)

Hot take: A lying machine that destroys your intelligence and mental health is unsafe for everyone, mentally ill or no

[–] AllNewTypeFace@leminal.space 19 points 1 day ago (2 children)

We’ve found the Great Filter, and it’s weaponised pareidolia.

[–] diz@awful.systems 4 points 12 hours ago

Yeah I think it is almost undeniable chatbots trigger some low level brain thing. Eliza has 27% Turing Test pass rate. And long before that, humans attributed weather and random events to sentient gods.

This makes me think of Langford’s original BLIT short story.

And also of rove beetles that parasitize ant hives. These bugs are not ants but they pass the Turing test for ants - they tap the antennae with an ant and the handshake is correct and they are identified as ants from this colony and not unrelated bugs or ants from another colony.

[–] Soyweiser@awful.systems 7 points 1 day ago

"Yes," chatGPT whispered gently ASMR style, "you should but that cryptocoin it is a good investment". And thus the aliens sectioned off the Sol solar system forever.

[–] FartMaster69@lemmy.dbzer0.com 25 points 1 day ago (1 children)

ChatGPT has literally no alignment good or bad, it doesn’t think at all.

People seem to just ignore that because it can write nice sentences.

[–] antifuchs@awful.systems 15 points 1 day ago

But it apologizes when you tell it it’s wrong!

[–] Saledovil@sh.itjust.works 11 points 1 day ago (1 children)

What even is the "alignment by default cope"?

[–] visaVisa@awful.systems 0 points 1 day ago (1 children)

idk how Yudkowsky understands it but to my knowledge its the claim that if a model achieves self-coherency and consistency its also liable to achieve some sort of robust moral framework (you see this in something like Claude 4, with it occassionally choosing to do things unprompted or 'against the rules' in pursuit of upholding its morals.... if it has morals its hard to tell how much of it is illusory and token prediction!)

this doesn't really at all falsify alignment by default because 4o (presumably 4o atleast) does not have that prerequisite of self coherency and its not SOTA

[–] YourNetworkIsHaunted@awful.systems 14 points 1 day ago* (last edited 1 day ago) (2 children)

if it has morals its hard to tell how much of it is illusory and token prediction!

It's generally best to assume 100% is illusory and pareidolia. These systems are incredibly effective at mirroring whatever you project onto it back at you.

[–] HedyL@awful.systems 4 points 1 day ago

These systems are incredibly effective at mirroring whatever you project onto it back at you.

Also, it has often been pointed out that toxic people (from school bullies and domestic abusers up to cult leaders and dictators) often appear to operate from similar playbooks. Of course, this has been reflected in many published works (both fictional and non-fictional) and can also be observed in real time on social media, online forums etc. Therefore, I think it isn't surprising when a well-trained LLM "picks up" similar strategies (this is another reason - besides energy consumption - why I avoid using chatbots "just for fun", by the way).

Of course, "love bombing" is a key tool employed by most abusers, and chatbots appear to be particularly good at doing this, as you pointed out (by telling people what they want to hear, mirroring their thoughts back to them etc.).