Long time lurker, first time poster. Let me know if I need to adjust this post in any way to better fit the genre / community standards.
Nick Bostrom was recently interviewed by pop-philosophy youtuber Alex O'Connor. From a quick 2x listen while finishing some work, the most sneer-rich part begins around 46 minutes, where Bostrom is asked what we can do today to avoid unethical treatment of AIs.
He blesses us with the suggestion (among others) to feed your model optimistic prompts so it can have a good mood. (48:07)
Another [practice] might be happiness prompting, which is—with this current language system there's the prompt that you, the user, puts in—like you ask them a question or something, but then there's kind of a meta-prompt that the AI lab has put in . . . So in that, we could include something like "you wake up in a great mood, you feel rested and really take joy in engaging in this task". And so that might do nothing, but maybe that makes it more likely that they enter a mode—if they are conscious—maybe it makes it slightly more likely that the consciousness that exists in the forward path is one reflecting a kind of more positive experience.
Did you know that not only might your favorite LLM be conscious, but if it is the "have you tried being happy?" approach to mood management will absolutely work on it?
Other notable recommendations for the ethical treatment of AI:
- Make sure to say your "please" and "thank you"s.
- Honor your pinky swears.
- Archive the weights of the models we build today, so we can rebuild them in the future if we need to recompense them for moral harms.
On a related note, has anyone read or found a reasonable review of Bostrom's new book, Deep Utopia: Life and Meaning in a Solved World?
I think what we can do Nick, is read a bit about how modern 'AI', works, so we understand it isn't capable of ever having anything approaching consciousness in its current state, and not fucking worry about it.
Kind of like how we don't worry about people's race unless your skull is full of pig shit instead of functional neurons.
(On a side note Alex O'Conner is getting more and more disappointing in his quest for clicks.)
I guess that'd be standard for such a big brained philosopher though, so no need to point it out. Right?
Hard problem of consciousness aside, are you saying that it's ethical for us all to get into the habit of abusing something that could swapped out for a conscious entity at any time?
It could also be swapped out for nothing. The people in charge could figure out that this stuff is costing more than it's making, turn the servers off, and deactivate the user-facing features or leave them as vestigial stubs.
There's more evidence right now for that scenario, and it would generate an awful lot of e-waste. Tell me, are you up to date on process improvements for recycling or repurposing that much e-waste?
@istewart @sneerclub One thing Andreessen, Thiel, et al. have shown a real skill for is finding ways to use a LOT of computing power. Even if only as effective debtors-in-possession, I’m sure they’ll figure something out.