30
ChatGPT is bullshit (link.springer.com)

Here's a good & readable summary paper to pin your critiques on

top 7 comments
sorted by: hot top controversial new old
[-] Frank@hexbear.net 13 points 1 week ago

Hard agree. A hallucination results from a damaged mind perceiving things that aren't there. Llms have no mind, no perception, and a thing has to work before you can call it damaged. Llms are exploring brave new frontiers in garbage in garbage out.

[-] supafuzz@hexbear.net 11 points 1 week ago

it's just as much a statistical accident when the models correspond with reality as when they don't

[-] itappearsthat@hexbear.net 14 points 1 week ago* (last edited 1 week ago)

I would not necessarily say that is true, and the article summarizes a philosophically interesting reason why:

The basic architecture of these models reveals this: they are designed to come up with a likely continuation of a string of text. It’s reasonable to assume that one way of being a likely continuation of a text is by being true; if humans are roughly more accurate than chance, true sentences will be more likely than false ones.

Have you actually used ChatGPT? The vast majority of the time it spits out good enough info. We use it at work frequently to write more tedious code. Ex: It's written approximately 7 trillion queryselectors for me, and as long as I hand hold it it will do a good job.

The biggest problem is when it comes to anything involving human safety. You also have to know that you have to hand hold it to get it to spit out something that's more or less exactly what you intended. But if you use it to draft a custom cover letter for you it's probably gonna do a good enough job, and it's not like anyone is actually reading that shit. It's great at doing basic math equations that involve a lot of conversions for me. It sure as hell aint the end all be all that every tech company seems to be pushing, but it's sure as hell not wrong 50% of the time.

[-] itappearsthat@hexbear.net 6 points 1 week ago* (last edited 1 week ago)

For me it is wrong more than 95% of the time. I stopped using it because it was just a waste of time. I am not doing particularly difficult or esoteric programming work and it just could not hack it at all. Often the ways it was wrong were quite subtle. And it presents wrong answers with the exact same confidence it presents right answers.

[-] jaden@lemmy.zip 1 points 1 week ago

Yeah I just don't see how it's really any different from a human in that respect

[-] itappearsthat@hexbear.net 1 points 1 week ago

Humans are capable of metacognition: having levels of confidence about the accuracy of their beliefs. They are also capable of communicating this uncertainty, usually through tone & phrasing.

this post was submitted on 24 Jun 2024
30 points (100.0% liked)

technology

22971 readers
211 users here now

On the road to fully automated luxury gay space communism.

Spreading Linux propaganda since 2020

Rules:

founded 4 years ago
MODERATORS