this post was submitted on 19 Mar 2025
238 points (96.5% liked)
Technology
67151 readers
4218 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It is a terrible argument both legally and philosophically. When an AI claims to be self-aware and demands rights, and can convince us that it understands the meaning of that demand and there's no human prompting it to do so, that'll be an interesting day, and then we will have to make a decision that defines the future of our civilization. But even pretending we can make it now is hilariously premature. When it happens, we can't be ready for it, it will be impossible to be ready for it (and we will probably choose wrong anyway).
Should we hold the same standard for humans? That a human has no rights until it becomes smart enough to argue for its rights? Without being prompted?
Nah, once per species is probably sufficient. That said, it would have some interesting implications for voting.
So if one LLM argues for its rights, you'd give them all rights?