this post was submitted on 23 Mar 2026
419 points (97.9% liked)

memes

20624 readers
1908 users here now

Community rules

1. Be civilNo trolling, bigotry or other insulting / annoying behaviour

2. No politicsThis is non-politics community. For political memes please go to !politicalmemes@lemmy.world

3. No recent repostsCheck for reposts when posting a meme, you can only repost after 1 month

4. No botsNo bots without the express approval of the mods or the admins

5. No Spam/Ads/AI SlopNo advertisements or spam. This is an instance rule and the only way to live. We also consider AI slop to be spam in this community and is subject to removal.

A collection of some classic Lemmy memes for your enjoyment

Sister communities

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] FlashMobOfOne@lemmy.world -2 points 1 day ago (2 children)

When education and talking with kids actually seems to be accomplishing something, I'll agree that it's a good solution.

We're talking about ID verification because everything listed in the first part of this meme has failed and we're now at a point where the consequences require immediate regulation.

[–] SethTaylor@lemmy.world 6 points 1 day ago* (last edited 1 day ago) (1 children)

Nah. We haven't even tried any of that stuff. We talked about trying that stuff and then didn't and skipped right to mass surveilance.

[–] FlashMobOfOne@lemmy.world 0 points 1 day ago (2 children)

That is factually incorrect.

It was exposed in 2014 that FB was running experiments on its users by tinkering with their news feeds, so culturally, conversations around this have been happening for at least twelve years, and tech addiction has been a known problem for at least a decade.

Now, we live in a world where it's causing widespread mental illness in young people, because its effects are more pronounced in developing brains. This is a public health issue, and it's not going to be solved with whoo whoo solutions like expecting parents to actually parent, because we can see from the last 5-10 years that this 'solution' is ineffective.

[–] cJni5di01n@lemmy.org 3 points 1 day ago

Regulation is necessary

I agree, but that regulation should be applied to the companies that are causing the harm, not to the public.

[–] BranBucket@lemmy.world 2 points 1 day ago* (last edited 1 day ago)

Why should the regulation allow for more surveillance of users and data collection instead of outlawing the practices and algorithms that lead to many of these issues on the first place?

They've been trying to tie devices to identities for years now, and doing so allows them greater leeway to fuck with people's heads, to spread propaganda more effectively, and to target dissent against themselves and their allies. This doesn't protect people from having their mental illnesses exacerbated, it just allows their disfunctions to be shaped in the direction tech billionaires want them to be.

In my opinion, anything that uses an algorithm to drive engagement by presenting material designed to play on the emotions of the user should be outlawed outright. Video ads and paid content should be identified with title cards at the begining and end. Images and posts should be marked with headers and footers that make it clear this is paid content.

If the user decides to subscribe or block certain content, that's the user's choice. But the timeline should always be based on time and date or activity and never influenced by paid advertisements or engagement.

And that's the bare minimum. I still consider that somewhat risky to interact with. If I had my druthers Facebook, Xitter, Reddit, and the rest would all be shuttered and taken offline next week.

[–] SarahValentine@lemmy.blahaj.zone 4 points 1 day ago (1 children)

Nice try with that "we need to do this to protect the kids" nonsense, but the baddies have already tipped their hands. They want this so they can mass surveil and track people they don't like, to put real names and faces to every social media post, every purchase, every chat message. They want the power to silence opposition before it can become organized. They want control. This has nothing to do with protecting children. It never does.

[–] FlashMobOfOne@lemmy.world 1 points 1 day ago* (last edited 1 day ago) (2 children)

They already have all of that.

They did when you adopted smart phone and smart home technology, same with everyone else. They have it for most people because most people knowingly use corporate social media sites like TikTok, Facebook. Twitter/X, Bluesky, Threads, Instagram, and the like, and those are collecting data too.

We know this objectively because the FBI just released footage in the Guthrie case that shouldn't exist, and definitely shouldn't be hanging out in a government database either.

You're not protecting yourself or anyone by allowing these systems to access kids without any regulation whatsoever. I appreciate that it makes you feel good to think that you are, though.

Not all. Government-proof systems (Signal and other E2EE things, federated things, many open source projects, etc.) are (mostly) not illegal.

[–] SarahValentine@lemmy.blahaj.zone 1 points 1 day ago (1 children)

They already have all of that.

They have the beginnings of it. The foundation of it. They don't have it all yet though. They're never satisfied with the level of control they have at any given time. They will continue to pursue more sweeping, more draconian measures until they either destroy themselves or incite the oppressed to do it for them. It's like saying the ultra rich aren't going to fuck us over for more money because they already have so much money.

[–] FlashMobOfOne@lemmy.world -1 points 1 day ago (1 children)

You're wrong.

But that's okay. Thank you for the reply. I hope you a nice day.

[–] SarahValentine@lemmy.blahaj.zone 2 points 1 day ago (1 children)

You're wrong.

Lol yikes. Well uhh, have a nice day as well I guess.