this post was submitted on 04 Feb 2026
141 points (100.0% liked)

Fuck AI

5558 readers
1249 users here now

"We did it, Patrick! We made a technological breakthrough!"

A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.

AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] pixxelkick@lemmy.world -4 points 4 hours ago (2 children)

To be clear: this isnt an AI problem, the LLM is doing exactly what its being told to

This is an Openclaw problem with the platform itself doing very very stupid things with the LLM lol

We are hitting the point now where, tbh, LLMs are on their own in a glass box feeling pretty solid performance wise, still prone to hallucinating but the addition of the Model Context Protocol for tooling makes them way less prone to hallucinating, cuz they have the tooling now to sanity check themselves automatically, and/or check first and then tell you what they found.

IE a MCP to search wikipedia and report back with "I found this wiki article on your topic" or whatever.

The new problem now is platforms that "wrap" LLMs having a "garbage in, garbage out" problem, where they inject their "bespoke" stuff into the llm context to "help" but it actually makes the LLM act stupider.

Random example: Github Copilot agents get a "tokens used" thing quietly/secretly injected to them periodically, looks like every ~25k tokens or so

I dunno what the wording is they used, but it makes the LLM start hallucinating a concept of a "deadline" or "time constraint" and start trying to take shortcuts and justifying it with stuff like "given time constraints I wont do this job right"

Its kinda weird how such random stuff that seems innocuous and tries to help can actually make the LLM worse instead of better.

[–] Windex007@lemmy.world 7 points 2 hours ago (1 children)

You had me up until your first sentence.

[–] pixxelkick@lemmy.world 1 points 1 hour ago (1 children)

Everything I said was very much correct.

LLMs are fairly primitive tools, they arent super complex and they do exactly what they say they do.

The hard part is wrapping that up in an API that is actually readable for a human to interact with, because the lower level abstract data of what an LLM takes in and spits out arent useful for us.

And then even harder is wrapping THAT API in another one that makes the input/output USEFUL for a human to interact with

You have layers upon layers of abstraction overtop of the tool to make it go from just a bunch of raw float values a human wouldnt understand, to becoming a tool that does a thing

That "wrapper" is what one calls the "platform".

And making a platform that doesnt fuck it up is actually very very hard, and very very easy to get wrong. Even a small tweak to it can substantially shift how it works

Think of it a lot like an engine in a car. The LLM is the engine, which on its own is not actually super useful. You have to actually connect that engine to something to make it do anything useful.

And even just doing that isnt very useful if you cant control it, so we take the engine and wrap it up in a bunch of layers of stuff that allow a human to now control it and direct it.

But, turns out, when you put a V6 engine inside a car, even a tiny little bit of getting the engineering wrong can cause all sorts of problems with the engine and make it fail to start, or explode, or fall out of the car, or stall out, or break, or leak... and unlike car engines, these engines are very very new and most engineers are still only just now starting to break ground on learning how to control them well and steer them and stop them from tearing themselves out of the car, lol.

So, to bring this back to the original post:

Most LLMs (engines) are actually pretty good nowadays, but the problem was Clawdbot (a specific brand of car manufacturer) super fucked up the way they designed their car so the car itself had a very very stupid engineering mistake. IE in this case, the brakes didnt work well enough and the car drove off a cliff.

That has nothing to do with how good the engine is or is not, the engine was just doing its job. The problem was with some other part of the car entirely, the part of the car Clawdbot made that wraps around the engine.

[–] Windex007@lemmy.world 0 points 1 hour ago (1 children)

You keep asserting they do exactly what they say they do.

Who is "they"

[–] pixxelkick@lemmy.world 2 points 38 minutes ago

When using the word "they", in English it refers the the last primary subject you referred to, so you should be able to infer what "they" referred to in my sentences. I'll let you figure it out.

"I love wrenches, they are very handy tools", in this sentence, the last subject before the word "they" was "wrenches", so you should be able to infer that "they" referred to "wrenches" in that sentence.

[–] TropicalDingdong@lemmy.world 14 points 3 hours ago (1 children)

I don't think we've overcome the halfglass of wine issue, rather, we've papier-mâchéd over some fundamental flaws in precisely what it is happening when an LLM creates the appearance of reason. In doing saw we're baking a certain amount of sawdust into the cake, and the fact that no substantive advances has really been made since maybe the 4, 4.5 days, with most of the "improvements" being seen coming from basically better engineering, its clear we've hit an asymptote with what these models are capable/ will be capable, and it will never manifest into a full reasoning system that can self correct.

There is no amount of engineering sandblasting that can overcome issues which are fundamental to the models structure. If the rot is in the bones, its in the bones.

[–] pixxelkick@lemmy.world 1 points 1 hour ago (2 children)

Nah there have been huge advancements in the past few months, you are definitely out of touch if you havent witnessed them

Recent models have gotten WAY better at "second guessing" themselves, and not acting nearly so confidently wrong.

I don’t think we’ve overcome the halfglass of wine issue

That isnt an LLM issue at all, that has nothing to do with LLMs in fact. Thats a problem with Stable Diffusion which is an entirely different kind of AI, but yeah that issue is fundamental to what stable diffusion is.

with most of the “improvements” being seen coming from basically better engineering

I mean, thats not much different from any other tech, a LOT of advanced tech we have today is dozens and dozens of separate bits of engineering all working in tandem to create something more meaningful.

Your smartphone has countless different and distinct advancements on different types of technology that come together to make a useful device, and if you removed any one of those pieces from it, it would be substantially less useful as a tool.

So yeah, I personally will very much count the other pieces of the puzzle, advancing, as the system as a whole advancing.

LLMs today compared to ones a year ago are quite a bit better, by a large degree, and the tooling around them has also improved a lot. The proliferation of Model Context Protocol Tools is proving to be a massive part of the system as a whole becoming something actually very useful.

[–] jaredwhite@humansare.social 1 points 16 minutes ago

Perhaps you didn't notice the forum you're posting in. We're not here because we love hearing slopaganda.

Personally I believe MCP is the new AMP, and I look forward to dancing on its grave.

[–] TropicalDingdong@lemmy.world 2 points 59 minutes ago* (last edited 53 minutes ago) (1 children)

I'm not out of touch whatsoever. I'm in the cut, and I've been here since long before LSTM's, and even perceptrons. I can almost promise you I'm deeper into this world than you'll ever be. I publish on this stuff.

LLMs today compared to ones a year ago are quite a bit better, by a large degree

No. They aren't. They've stalled and its very clear they've stalled. There have been improvements in some of the background engineering that create the illusion of model improvement, but this is fundamentally different than the improvements we saw from the earliest transformers to gpt's, from 2021-2023/4.

That isnt an LLM issue at all, that has nothing to do with LLMs in fact.

No, it is. And there is no clear way around it. It is an LLM issue because its a transformers issue, and it might even go deeper and be a back prop issue.

[–] pixxelkick@lemmy.world 0 points 41 minutes ago (1 children)

The "wine glass half full" thing, I assume, is you referring to the problem surrounding trying to image generate a specific glass of wine, or similar issues of "generate a room that definitely doesnt have an elephant in it, its devoid of any elephants, zero elephants in the room"

This is specifically a stable diffusion problem, and doesnt really apply to LLMs in the same manner.

[–] TropicalDingdong@lemmy.world 3 points 36 minutes ago* (last edited 34 minutes ago)

Its not a problem specific to any model. Its present in all LLM's and possibly/ probably all transformers, and potentially even deeper. I get you don't get it, so just go take a break.

Not being able to generate something like a glass of wine is just a symptom of something far more significant.