this post was submitted on 28 Feb 2026
392 points (96.2% liked)

Technology

82070 readers
5689 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] sturmblast@lemmy.world 4 points 11 hours ago

My boss can't wrap his head around why handing me a direct printout of LLM output is not acceptable

[–] johncandy1812@lemmy.ca 13 points 17 hours ago (1 children)

I hate how AI is used to make deep fakes, revenge porn, CP - and people tolerate it because "they're working out the issues."

How about they work those out BEFORE they give people access to these tools.

[–] RememberTheApollo_@lemmy.world 3 points 6 hours ago

They tolerate it because it’s easy, they can copy-paste, and they need even less critical thought about the output than having to search for and choose what might be a viable source of decent information.

The issues aren’t bugs. They’re acceptable flaws in the search for investment capital.

[–] BitsAndBites@lemmy.world 7 points 16 hours ago (2 children)

My coworkers are doing this to me. They are even pasting into PR reviews. The threat is real.

[–] softwarist@programming.dev 4 points 12 hours ago

It's even better when they copy-paste slop answers that are flat out wrong without bothering to check.

[–] dwemthy@lemmy.world 4 points 14 hours ago

My org has our PRs reviewed by an AI automatically

[–] GreenBeard@lemmy.ca 40 points 1 day ago (1 children)

Absolutely rude. If you're using AI to make a point for you, you've already admitted you don't know enough about what you're talking about to be having a opinion in the first place, let alone be worth discussing an issue with.

[–] partofthevoice@lemmy.zip 24 points 21 hours ago* (last edited 21 hours ago) (3 children)

I’ve had these interactions with the head of my IT department. I asked to procure a license for jfrog artifactory. He literally copy/pasted a ChatGPT response to me that began like this:

Here’s a breakdown of how JFrog Artifactory compares to using GitHub, NPM, or other language-specific package mangers (like Pypi)…

1. Purpose and Functionality

2. Workflow & Developer Experience

3. Security and Compliance

When to use JFrog

It came with a bunch of theoretical risks that are completely resolved by the simple ability of just not being a complete fucking moron.

It was really frustrating that I tried to talk with my IT leader, and instead found a proxy for ChatGPT.

After that, he created a group chat with him, I, and my colleagues in security. He proceeded to paste ChatGPT output outlining bullshit risks and theories, with the implicit expectation that I rhetorically address each of them via my own response. I’d explain things like,

“[well if you read the fucking request yourself, you’d know that] we aren’t planning to use the software that way, so the concern isn’t relevant. Even if we were though, those problems are easily addressable via …”

In some cases, I even had to explain that the problems he’s raising are already problems faced in the current ecosystem. Completely unrelated to the software I’m talking about… ChatGPT just straight up implying that an architectural problem is a software risk.

I’d reply, and I swear to god he’d just give ChatGPT my text and paste the reply from ChatGPT back to me.

I lost a lot of respect for him. Why the fuck would you do that?

[–] jason@discuss.online 9 points 17 hours ago

That guy to all his friends: "AI makes me 10x more productive!"

[–] Panthenetrunner@lemmy.dbzer0.com 11 points 20 hours ago

I'm fast coming to the conclusion that AI can indeed replace jobs. The thing is that the only job it can actually replace is that of a lazy middle manager. AI is great at responding to email if A:) you don't know what your talking about or B:) you don't respect the other person enough to waste the time formulating an actual response. AI in my experience is only really good at faking that there's someone on the other end. The fact that there's an entire management class it can convenienceingly impersonate is a pretty searing indictment as far as I'm concerned.

[–] Natanael@infosec.pub 6 points 18 hours ago* (last edited 18 hours ago)

This gets at my own personal perspective of using LLMs to respond - it's not just about not putting effort into understanding and responding yourself, rather it is about making yourself a proxy to a tool I could use myself, and doing so *without even having a better understanding of how to use the tool to answer my question*, and still thinking you're somehow made a positive contribution, that is the most disrespectful.

If you genuinely thought the LLM could help me then you should be explaining your process to me for how to use it and validate responses, or else at least you should ask me for more info and explain how you think it's responses could help if you really do think you're better at operating it.

Imagine doing the same in a workshop, and taking a powertool to an object before you even bothered figuring out what the other person wanted. Or trying to be helpful by asking questions on your behalf to other departments, but messing up the context and thus repeatedly producing useless answers that you have to put time into refuting.

[–] aesthelete@lemmy.world 50 points 1 day ago* (last edited 23 hours ago) (1 children)

Totally agree. When someone sends me some AI slop about a topic I have knowledge about -- which I've had this happen to me recently during a debug session -- and asks me to read it, I think to myself "this person does not respect me, otherwise they wouldn't be telling me to read stuff that may or may not be accurate that they themselves never read." It's like a new, worse version of "let me google that for you" but without the sarcasm, and without the results actually being helpful.

[–] sockenklaus@sh.itjust.works 3 points 18 hours ago* (last edited 18 hours ago) (1 children)

I know that feeling. I experienced it more than one time in areas of law I consider myself a little bit more knowledgeable than the average person. It's just a slap to the face to try to discuss a topic that you know a little bit about with an AI.

The thing is: I am 100 % sure those people use LLM answer not out of disrespect but because they honestly believe that an LLM produces a better argument than they possibly could themselves.

[–] aesthelete@lemmy.world 1 points 12 hours ago

The thing is: I am 100 % sure those people use LLM answer not out of disrespect but because they honestly believe that an LLM produces a better argument than they possibly could themselves.

And I have zero confidence your 100% because you have zero backing for your claim other than believing people have good intentions.

[–] jason@discuss.online 28 points 1 day ago (1 children)

My company hired a consulting firm to help with a transition period. The consulting firm sent my boss an email that outlined the plans for what we should do and how they are going to help. Without directly giving it away, the email was clearly AI output, and my boss instantly terminated their contract. We aren't exactly anti-AI, but to the point of the post, it's just so rude... and my boss is pretty fuckin cool.

[–] mcv@lemmy.zip 17 points 22 hours ago

Especially rude if you want to charge money for it. If your boss wanted an AI answer, they would have asked an AI. You don't need an expensive consulting company for that.

[–] reksas@sopuli.xyz 3 points 16 hours ago (1 children)

my mother constantly keeps sending me texts that are just direct copy-paste from llm output. can't even tell her to stop doing it because she just ignores me if i say something she doesnt want to hear.

[–] JohnEdwa@sopuli.xyz 2 points 7 hours ago (1 children)

Ask chat GPT to come up with a nice message explaining why direct copy pastes of LLM outputs is bad. Copy paste it to her directly.
Maybe she will understand it better that way.

[–] reksas@sopuli.xyz 1 points 7 hours ago

no, she just think she is being helpful but doesnt care what i think about it because apparently she knows everything better. She would just ignore that or otherwise make me even more annoyed.

[–] benny@reddthat.com 2 points 16 hours ago* (last edited 16 hours ago)

Chat is just the wrong interface to AI, period. If you use it as an agentic tool with human review, it either works or doesn't and you can keep improving it for the task at hand.

[–] HerbalGamer@sh.itjust.works 1 points 14 hours ago

AI etiquette

╭─────────────────────────╮ │ I asked ChatGPT and │ │ here's what it told me: │ │ ... │ ╰─────────────────────────╯ ╭─────────╮ │ blocked │ ╰─────────╯

Time to learn AI manners

[–] MrPnut@lemmy.world 35 points 1 day ago (5 children)

Whenever someone at work says "ChatGPT says this" or "Claude says this" or "I asked Gemini and..." whatever they say after that point is just static and I never take them seriously as a person again.

[–] pHr34kY@lemmy.world 24 points 1 day ago (1 children)

I appreciate the honesty when they say it's an AI response and not genuine knowledge.

When I tell someone "an LLM told me that..." It's usually followed by "Let's see if there's any truth to it." An AI response should always be treated as a suggestion, not an answer.

Hell, Google's AI still doesn't know which day the F1 GP is on this week. It was wrong by a whole week a while back. Now it's only off by a day.

[–] mcv@lemmy.zip 12 points 22 hours ago (1 children)

An AI response should always be treated as a suggestion, not an answer

Exactly. An AI response can be a great way to get started on a topic you know little about, but it's never a definitive answer. You have to verify whether it's actually true. Whether it works. Never trust it blindly.

[–] Panthenetrunner@lemmy.dbzer0.com 4 points 20 hours ago (1 children)

I feel like a big barrier is people anthropomorphizing the AI. It's not "ChatGPT generated this" it's "ChatGPT said this". I don't necessarily blame people for it, machine that speaks to you short circuits something in people's brains and it's not like we've got better language to talk about it. It's just that... people treat it as an opinion, not as software output. And so long as that's how people handle it, I just don't know if a "healthy" use of the technology is possible.

[–] mcv@lemmy.zip 3 points 19 hours ago

Exactly. We are extremely social animals, hardwired to recognise ourselves in things around us, which I'm sure is super useful and vital for a tribe of hunter gatherers living in a hostile environment. But it means that now we recognise faces and emotions in power outlets and lawn chairs. It's really not surprising we see intelligence and awareness in LLMs, because we recognise that stuff in everything. We are really poor at the level of critical thought required to deal with this responsibly.

I never take them seriously as a person again

i dunno dude. i used to be a real piece of shit.

[–] vacuumflower@lemmy.sdf.org 4 points 21 hours ago

As a source it's rude. As a piece of unreliable help of the kind "we both don't know the syntax of that programming language, let's ask Ollama how to draw such and such a shape in it" it's kinda fine.

[–] pkjqpg1h@lemmy.zip 6 points 23 hours ago
load more comments (1 replies)
[–] pixxelkick@lemmy.world 110 points 1 day ago (6 children)

Something that some coworkers have started doing that is even more rude in my opinion, as a new social etiquette, is AI summarizing my own writing in response to me, or just outright copypasting my question to gpt and then pasting it back to me

Not even "I asked chatgpt and it said", they just dump it in the chat @ me

Sometimes I'll write up a 2~3 paragraph thought on something.

And then I'll get a ping 15min later and go take a look at what someone responded with annnd... it starts with "Here's a quick summary of what (pixxelkick) said! "

I find this horribly rude tbh, because:

  1. If I wanted to be AI summarized, I would do that myself damnit
  2. You just clogged up the chat with garbage
  3. like 70% of the time it misquotes me or gets my points wrong, which muddies the convo
  4. It's just kind of... dismissive? Like instead of just fucking read what I wrote (and I consider myself pretty good at conveying a point), they pump it through the automatic enshittifier without my permission/consent, and dump it straight into the chat as if this is now the talking point instead of my own post 1 comment up

I have had to very gently respond each time a person does this at work and state that I am perfectly able to AI summarize myself well on my own, and while I appreciate their attempt its... just coming across as wasting everyones time.

[–] MrKoyun@lemmy.world 9 points 1 day ago

I hate people so fucking much

[–] XLE@piefed.social 31 points 1 day ago (1 children)

This is sad, really. People are fed the lie that AI is objective, and apparently they think that they will get the objective summary of what you said if they run it through a chatbot.

And the more people interact with chatbots, the harder they find it to interact outside of the chatbots. So they might feel even more uncomfortable with asking you to summarize yourself. So they go back to the chatbot. It's a self-perpetuating cycle.

[–] ErmahgherdDavid@lemmy.dbzer0.com 12 points 20 hours ago* (last edited 20 hours ago)

Exactly. To your point, AI output is probabilistically the average opinion of everyone on the internet so it shares the common biases of the general public. Even with a bit of RLHF to "balance out" the models. Also it probably doesn't help to anthropomorphise them. They don't have opinions, they just autocomplete based on prior input

It seems pretty clear after a few years of people getting AI psychosis that LLMs are an addictive psychological hazard

load more comments (4 replies)
[–] Valmond@lemmy.dbzer0.com 3 points 20 hours ago

I asked chatgpt and it told me:

Wrong network configuration

load more comments
view more: next ›