So every AI’s gonna identify as an Arch user with striped socks now?
Fediverse
A community dedicated to fediverse news and discussion.
Fediverse is a portmanteau of "federation" and "universe".
Getting started on Fediverse;
- What is the fediverse?
- Fediverse Platforms
- How to run your own community
Ahahahahaha, so it's going to be a self-hating Meta AI bot?
lemmygrad
imagining Zuck launching his "everybody gets ten virtual friends" initiative and accidentally re-radicalizing your parents and grandparents in the other direction.
The bot trained on hexbear and lemmygrad vs the bot trained on .world:
Damn zuckbot's gonna end up being a commie-bot that posts absurdist memes about beans if it's harvesting hexbear posts for content
The AI wasting hours of processing power having an internal struggle session re: outdoor cats before simply replying with ":pigpoopballs" on a platform that doesn't have that emoji
Going straight to palantir
now I feel I should upload my asshole pic.
I'll be upping my use of Maoist Standard English and in response this revelation.
Peertube as well. 46 instances.
Oh and https://mastodon.sdf.org/ as well.
Sopuli's there also! This sucks, but hopefully Anubis protects against Meta.
Honestly, I already figured my posts probably were being used to train a LLM without my consent.
I'm more concerned about the non-consensual scraping causing excess load on the servers. The taking of content without license to train their energy-wasting autocomplete that is being used to for little commercially but to try to cheapen labor and pocket the money is a problem too. But I hate having servers impacted by their bullshit.
Check out the robots.txt on any Lemmy instance....
Linked article in the body suggests that likely wouldn't have made a difference anyway
The scrapers ignored common web protocols that site owners use to block automated scraping, including “robots.txt” which is a text file placed on websites aimed at preventing the indexing of context
Yeah ive seen the argument in blog posts that since they are not search engines they dont need to respect robots.txt. Its really stupid.
"No no guys you don't understand, robots.txt actually means just search engines, it totally doesn't imply all automated systems!!!"
If they have a brain, and they do have the experience from Threads, they don't need to scrape Lemmy. They can just set up a shell instance, subscribe to Lemmy communities, and then use federation to get their data for free. That doesn't use robots.txt at all regardless.
Oh that's certainly a decision they made
aussie.zone and beehaw.org are on the list as well
I thought we all knew and were training it wrong on purpose..
...as a joke.
Aw hell nah
Our cdn is there... Joy...