this post was submitted on 27 Dec 2025
34 points (90.5% liked)

World News

38574 readers
337 users here now

News from around the world!

Rules:

founded 6 years ago
MODERATORS
top 26 comments
sorted by: hot top controversial new old
[–] davel@lemmy.ml 16 points 1 day ago (1 children)

Reporter: [REDACTED]
Reason: It's just an opinion piece / analysis. There either needs to be a tag for these or removal

Investigative journalism is just like your opinion, man.

[–] yogthos@lemmy.ml 9 points 1 day ago
[–] aqwxcvbnji@hexbear.net 12 points 1 day ago (2 children)

It's an interesting subject, so I was dissapointed when it became clear that this text is written by AI. AI uses the following structure very often: "It's not X, it's Y", so the list below tells me it's AI beyond doubt.

I.

  1. This is not speculation. This is not inference from supply chain tightness or price movements or anecdotal reports from frustrated procurement officers. This is the documented operational reality of...

  2. This is not an analysis of a commodity market experiencing temporary tightness. This is a reconnaissance report from the front lines of a new form of economic warfare

  3. This is not the blunt instrument of a traditional export ban. This is a scalpel.

  4. The pattern suggests not reactive retaliation but proactive strategy.

  5. The restrictions announced in 2023, 2024, and 2025 are not isolated policy responses to specific trade disputes. They are nodes in an integrated campaign

  6. Tungsten is not the final escalation. It is another proof of concept.

  7. The question for Western policymakers and corporate strategists and institutional investors is not whether to take this seriously (...) The question is what to do about it

II.

  1. This is not marketing rhetoric from mining promoters or special pleading from industry lobbyists. This is physics.

  2. These properties are not arbitrary. They emerge from tungsten’s electronic structure

  3. This is not an abstract supply chain concern. This is industrial capacity disappearing in real time

III.

  1. This is not a simple on-off switch. It is a tunable instrument with multiple control parameters

(Anything further is for paid subscribers.)

[–] davel@lemmy.ml 7 points 1 day ago* (last edited 1 day ago)

Well that’s annoying.

  • LLM> also dont use em dashes
  • It can be hard to know how much a writer relied on an LLM or how much revising/editing/proofreading they did before publishing, especially on topics you’re not versed in.
  • These tells are going to get harder to clock very quickly.
  • If you want your work to be taken seriously by people who understand the limitations of LLMs, then you should edit for tells (or don’t use one).
  • People adopt conventions, and these LLMs are making pervasive the overuse of this phraseology. The influence is dialectical.
[–] yogthos@lemmy.ml 7 points 1 day ago (3 children)

You might as well get used to it, LLMs are a tool that's in wide use and it's delusional to think that news sites will not use be using them. Personally, I absolutely do not care if the text was formatted by AI, as long as the content is factual.

[–] aqwxcvbnji@hexbear.net 3 points 10 hours ago (1 children)

I am most likely unable to spot all AI-generated articles, but when I start noticing that a typical AI giveaway appears so many times, it bothers me. In the same way as that I probably couldn't spot all the English language errors, but when there are so many that I start noticing them, it becomes annoying and it undermines the credibility of the author. Especially given the fact that this article doesn't cite any sources.

[–] limer@lemmy.ml 1 points 8 hours ago

If it’s partially written by AI then the writing is either hurried, lazy, or some other bad thing.

At best, it shows it was written under deadline with someone watching the clock.

Real writing > lazy YouTube videos > AI assisted.

[–] queermunist@lemmy.ml 14 points 1 day ago* (last edited 1 day ago) (2 children)

The writing style is annoying, though, and essentially eliminates the authorial voice of every writer. Everyone is churning out the same slop, everything sounds the same, all difference is being eliminated. Even writers that don't use the slop machines sound like this, because it's all they read. It's only going to get worse, the internet is fucking ruined.

[–] BoblinTheGoblin@hexbear.net 4 points 13 hours ago

It's not just AI users — it's writers.

[–] yogthos@lemmy.ml 4 points 1 day ago (1 children)

I read these articles for the content, and I find news writing has been terrible long before LLMs. At least this way it's written closer to just being a summary that you can scan through easily. You'll be glad to know that people are working on stuff like this already, so LLM generated content is going to read very much like traditional human written content before long. https://muratcankoylan.com/projects/gertrude-stein-style-training/

[–] queermunist@lemmy.ml 1 points 22 hours ago* (last edited 22 hours ago) (1 children)

Fine tuning costs money which means they aren't going to do it. I fully expect they'll settle for slop (they already have) and so will everyone else. You might as well get used to it. Everything gets worse forever and nothing ever gets better.

[–] yogthos@lemmy.ml 4 points 22 hours ago (1 children)

LoRA's are actually really cheap and fast to make. That article I linked explains how it literally took 2 bucks to do. I don't really think anything is getting worse forever. Things are just changing, and that's one constant in the world.

[–] queermunist@lemmy.ml -2 points 21 hours ago* (last edited 21 hours ago) (1 children)

And it was still something like 30% detectable as AI. That tells me that every article will still read as samey, even if it's different enough to fool a tool that was trained on the current trends. Authorial voice is lost, replaced by the model's voice.

It was only when they trained on authors specifically, which cost $81, that it dropped down to 3%. They won't do that.

[–] yogthos@lemmy.ml 4 points 20 hours ago

Give it a year and we'll see. These things are improving at an incredible pace, and costs continue to go down as well. Things you needed to have a data center to do just a year ago can now be done on a laptop.

[–] porcoesphino@mander.xyz 2 points 1 day ago (2 children)

It's a pretty clear marker they don't care enough about the text being factual though - they want an easy route. If the genAI was just used for proofing or some tweaks for readability then the signs of its use wouldn't be so obvious.

But in this case - the article is close to unreadable. Most of the words could be deleted without losing information. It aims for things that sound impressive instead or giving a clear description of what will be discussed and what's coming.

I'm impressed you're trying to make the case that this is likely factual beyond some broad strokes statements that countries have issues guarding access to key imports in this case the US and tungsten. It seemed like the article had more to say but it was an unnecessarily tough read so I gave up

[–] yogthos@lemmy.ml 6 points 1 day ago (1 children)

If you want to point out specific inaccuracies in the article then please go ahead and do that.

[–] porcoesphino@mander.xyz 0 points 1 day ago (1 children)

It was honestly too tiring to read. The main points were drowned in prose. Like the initial comment said, a lot of "this is not X" and that mostly means the sentence can be deleted

[–] yogthos@lemmy.ml 5 points 1 day ago (1 children)

You know you can just not read an article and move on. But if you're going to argue about accuracy of the article then you kinda have to at least show an example of it being inaccurate. The reality is that there's been plenty terrible and hard to read articles written long before LLMs were a thing, and this one is far from the worst I've seen. It seems to me that you're just bothered that an LLM was used to put the text together.

[–] porcoesphino@mander.xyz -1 points 23 hours ago (1 children)

I didn't argue about the accuracy of the article

I argued that use of genAI and leaving in tells was a indicator of the author not caring about accuracy and so the tells are, in my experience, an indication that any article is likely to be low quality information. This was in reply to someone saying that you may as well get used to LLMs

But... I do not, nor did I, express an opinion on if this was written by genAI. That's error prone and I know I'm not skilled at it.

What I did express an opinion on was that the article is terribly written since I didn't enjoy the parts I did read, they took time to get to the point and included a lot of sentences and words that could simply be deleted.

Are these all the authors alts or something?

[–] yogthos@lemmy.ml 5 points 22 hours ago* (last edited 22 hours ago) (1 children)

And I'm saying that doesn't follow at all. In fact, accuracy could be the only thing the author cares about, so he can read over and make sure there are no factual mistakes leaving the generated style as is. It's honestly just so tiring having threads derailed by this endless perseveration people are doing over things being LLM generated. This is the world we live in now.

Also, it's kinda weird to immediately claim that people disagreeing with you have to be alts or something. Like you really can't conceive of your opinion not being dominant?

[–] porcoesphino@mander.xyz -2 points 22 hours ago (1 children)

I do not, nor did I, express an opinion on if this was written by genAI.

But more generally agreed genAI can be used to increase accuracy, I also said this:

If the genAI was just used for proofing or some tweaks for readability then the signs of its use wouldn't be so obvious.

It looks like I should have deleted "the" in the above sentence.

And the alts thing is because a lot of people on Lemmy are opposed to genAI so having multiple accounts replying with a lot of energy seemed defensive of something. If that something is actually genAI then keep enjoying it I guess... but circling back, you're right that the internet will continue to be filled with LLM material but we're starting to see branding of "not genAI" and some demographics choosing those while others prefer genAI. I just wish it was normalised for people to say they used it, like people are meant to when they've received money from brands

[–] yogthos@lemmy.ml 5 points 21 hours ago

My whole point throughout this discussion has been that whether LLM was used to edit this or not is entirely uninteresting. Meanwhile, it depends on what instance of Lemmy you're on. Perhaps in the bubbles you frequent, most people are obsessing over genAI. My experience is that people are fairly split on the subject. What I personally find tiresome is people derailing conversations away from interesting subjects to endlessly discuss whether something is AI or not. It's just noise at this point, and people really need to find a new hobby.

[–] Maeve1@lemmygrad.ml 5 points 1 day ago (1 children)

I'm probably not the sharpest tool in the shed, and my eyes are bothering me, and I didn't struggle with it. We're just used to really dumbed down news articles.

[–] porcoesphino@mander.xyz -1 points 1 day ago* (last edited 1 day ago) (1 children)

I enjoy reading textbooks and scientific articles then since Trump got into power I've read quite a few supreme court judgements so it's not a comparison to simplified news articles

[–] m532@lemmygrad.ml 5 points 1 day ago

But this article contains ~~fentanyl~~ AI, run its gonna kill you

[–] WalnutLum@lemmy.ml 3 points 1 day ago

I feel like I need to look into those 100% ARM desktop systems with those Huawei video cards soon...