317

The logical end of the 'Solution to bad speech is better speech' has arrived in the age of state-sponsored social media propaganda bots versus AI-driven bots arguing back

top 50 comments
sorted by: hot top controversial new old
[-] ATQ@lemm.ee 86 points 1 year ago

Shit. I could have told them to just block lemmygrad for like $100 😂🤣😂

[-] Luci@lemmy.ca 23 points 1 year ago
[-] thefartographer@lemm.ee 6 points 1 year ago

Sold! $190 to watch to two of y'all fight Russian and Nazi sympathizers. I'm selling this on pay-per-view

[-] Swim@lemmy.ca 2 points 1 year ago* (last edited 1 year ago)

i would consider tuning into that, what you charging for a front seat to the action?

[-] thefartographer@lemm.ee 3 points 1 year ago
[-] Railcar8095@lemm.ee 3 points 1 year ago

Best I can do is an upvote.

Please tell me how to block a full instance

[-] mojo@lemm.ee 64 points 1 year ago

Just a reminder, LLMs are not designed to provide truth, but rather naturally sounding word generation.

[-] tehmics@lemmy.world 3 points 1 year ago

We can certainly argue over what they're designed to do, and I definitely agree that's the goal of them. The reality though is that on some level it is impossible to separate assertions from the words that describe them. Language itself is designed to communicate ideas, you can't really create language without also communicating ideas, otherwise every sentence from an LLM would just look like

"Has Anyone Really Been Far Even as Decided to Use Even Go Want to do Look More Like"

They will readily cite information that was fed to them. Sometimes it is on point, sometimes not. That starts to be a bit of an ethical discussion on whether it is okay for them to paraphrase information they were fed, and without citing it as a source of the info.

In a perfect world we should be able to expand a whole learning tree to trace back how the model pieced together each word and point of data it is citing, kind of like an advanced Wikipedia article. Then you could take the typical synopsis that the model provides and dig into it to judge for yourself if it's accurate or not. From a research standpoint I view info you collect from a language model as a step down from a secondary source and we should be able to easily see how it gets to that info.

load more comments (1 replies)
[-] Gradually_Adjusting@lemmy.world 40 points 1 year ago

After WWII in Germany, the cool young people knew you couldn't trust anyone over 30.

Nowadays, cool people need to understand that you can't trust anything bland and sanitized-sounding on the internet. For the rest of our lives, your personhood is on trial with everything you say.

It could tear society apart before we even know it's happening.

[-] theodewere@kbin.social 9 points 1 year ago

For the rest of our lives, your personhood is on trial with everything you say.

bravo, man really well said

[-] etuomaala@sopuli.xyz 4 points 1 year ago

This was why I was so furious about Elon Mask's blue checkmark debacle. He had a chance to prove that a gigantic part of the internet was a) human and b) non-duplicate. I was really shocked by how badly an apparently smart person fucked it up. Not so smart, it turns out.

[-] Gradually_Adjusting@lemmy.world 4 points 1 year ago

EM loses the ability to infuriate you when you understand him as a narcissist.

[-] aaaa@lemmy.world 2 points 1 year ago

Nowadays, cool people need to understand that you can’t trust anything bland and sanitized-sounding on the internet.

This is bad news for my communication style.

[-] Gradually_Adjusting@lemmy.world 2 points 1 year ago

Same but kinda not same

load more comments (1 replies)
[-] zephyreks@lemmy.ca 30 points 1 year ago* (last edited 1 year ago)

Ah yes, American truths like "Iraq has WMDs and that's why invading them is the fair and just thing to do," "abortion is bad for human rights," "the US isn't collecting all of your internet traffic because that would be a violation of privacy," and "this CIA-funded coup of a democratically-elected government will definitely help spread democracy around the world."

This researcher has built a pro-America AI disinformation machine for $400. I expect that, like most American media, it will start citing "independent think tanks" like Atlantic Council (which, coincidentally, is staffed mostly by ex-US intelligence and receives funding from US intelligence agencies) and use reports gathered by "independent sources" such as the US 4th PsyOps Airborne (which, per their recent recruiting videos, admits to orchestrating large-scale protests including Euromaidan, Tiananmen Square, and others).

[-] mea_rah@lemmy.world 12 points 1 year ago

Have you seen any tweet this bot generated that would contain misinformation? Because I haven't.

What is the context for Iraq WMDs? I haven't seen it anywhere in the article?

[-] zephyreks@lemmy.ca 4 points 1 year ago

Is anyone arguing that, at the time of the Iraq War, it wasn't considered a "truth" in America that Iraq was developing WMDs and that anything to the contrary was considered disinformation?

load more comments (2 replies)
load more comments (16 replies)
[-] etuomaala@sopuli.xyz 7 points 1 year ago

What makes you say that this new disinformation machine is pro-America?

[-] mayo@lemmy.today 14 points 1 year ago
[-] mob@lemmy.world 2 points 1 year ago

That's way worse than I imagined. Like 400$ seems like to much money spent

[-] autotldr@lemmings.world 7 points 1 year ago

This is the best summary I could come up with:


Russian criticism of the US is far from unusual, but CounterCloud’s material pushing back was: The tweets, the articles, and even the journalists and news sites were crafted entirely by artificial intelligence algorithms, according to the person behind the project, who goes by the name Nea Paw and says it is designed to highlight the danger of mass-produced AI disinformation.

Mitigations are possible, such as educating users to be watchful for manipulative AI-generated content, making generative AI systems try to block misuse, or equipping browsers with AI-detection tools.

In recent years, disinformation researchers have warned that AI language models could be used to craft highly personalized propaganda campaigns, and to power social media accounts that interact with users in sophisticated ways.

Renee DiResta, technical research manager for the Stanford Internet Observatory, which tracks information campaigns, says the articles and journalist profiles generated as part of the CounterCloud project are fairly convincing.

“In addition to government actors, social media management agencies and mercenaries who offer influence operations services will no doubt pick up these tools and incorporate them into their workflows,” DiResta says.

The CEO of OpenAI, Sam Altman, said in a Tweet last month that he is concerned that his company’s artificial intelligence could be used to create tailored, automated disinformation on a massive scale.


The original article contains 806 words, the summary contains 215 words. Saved 73%. I'm a bot and I'm open source!

[-] JohnDClay@sh.itjust.works 7 points 1 year ago

So is it against Russian disinformation, or is does it make anti Russia disinformation? I'd hope the former, it's easy enough to refute Russia with correct information.

[-] Draghetta@sh.itjust.works 8 points 1 year ago

I know it’s taboo but hear me out - you could read the article and find out

[-] MrBusinessMan@lemm.ee 4 points 1 year ago

Per the article, it’s the latter.

The tweets, the articles, and even the journalists and news sites were crafted entirely by artificial intelligence algorithms, according to the person behind the project, who goes by the name Nea Paw and says it is designed to highlight the danger of mass-produced AI disinformation.

[-] xkforce@lemmy.world 6 points 1 year ago

OpenAI is so concerned that AI will do x and y bad thing but still pour all these resources into developing it further.

[-] Spzi@lemm.ee 8 points 1 year ago

There are other endeavors where a great deal of the effort is put into making it safe. Space travel for example.

I wish that was the case for AI development. AI safety is a notoriously underfunded, understaffed and still overall neglected field.

[-] Touching_Grass@lemmy.world 3 points 1 year ago

OpenAI isn't responsible for what Russians do with it anymore than any company is for how users use their product

[-] xkforce@lemmy.world 8 points 1 year ago* (last edited 1 year ago)

If someone knows that what they're about to create is going to do harm like this, they shoulder some of the responsibility for those consequences. They dont just get to wash their hands of it as if they had no idea.

load more comments (10 replies)
[-] etuomaala@sopuli.xyz 3 points 1 year ago

In that case, would you object to the posting of detailed schematics on the internet for the creation of nuclear weapons?

load more comments (1 replies)
[-] jana@leminal.space 3 points 1 year ago

That concern is feigned, for PR.

load more comments (1 replies)
[-] Ertebolle@kbin.social 6 points 1 year ago

So this is why Elon is suddenly more upset than usual about bots

[-] Wahots@pawb.social 5 points 1 year ago

The Federal Election Commission has said it may limit the use of deepfakes in political ads.

Any use of deepfakes should serve as immediate disqualification/termination for any political candidate, and any donations immediately reversed.

load more comments (1 replies)
[-] chaircat 3 points 1 year ago

Honestly, if you look at it in a vacuum, this looks pretty similar to what the other side is doing.

It's a bot that draws from its own side's narratives and pushes that line.

Take away Russia from the picture and think about how often our media pushes a spin on other subjects that isn't exactly the truth.

Doesn't look so much like "social media propaganda bots versus AI-driven bots arguing back" as much as propaganda bots on both sides spewing whatever their masters want us to see.

load more comments (16 replies)
[-] Buelldozer@lemmy.today 2 points 1 year ago

Only $400? Where can I buy one?

load more comments (1 replies)
load more comments
view more: next ›
this post was submitted on 18 Sep 2023
317 points (95.7% liked)

World News

39364 readers
2207 users here now

A community for discussing events around the World

Rules:

Similarly, if you see posts along these lines, do not engage. Report them, block them, and live a happier life than they do. We see too many slapfights that boil down to "Mom! He's bugging me!" and "I'm not touching you!" Going forward, slapfights will result in removed comments and temp bans to cool off.

We ask that the users report any comment or post that violate the rules, to use critical thinking when reading, posting or commenting. Users that post off-topic spam, advocate violence, have multiple comments or posts removed, weaponize reports or violate the code of conduct will be banned.

All posts and comments will be reviewed on a case-by-case basis. This means that some content that violates the rules may be allowed, while other content that does not violate the rules may be removed. The moderators retain the right to remove any content and ban users.


Lemmy World Partners

News !news@lemmy.world

Politics !politics@lemmy.world

World Politics !globalpolitics@lemmy.world


Recommendations

For Firefox users, there is media bias / propaganda / fact check plugin.

https://addons.mozilla.org/en-US/firefox/addon/media-bias-fact-check/

founded 2 years ago
MODERATORS