A machine learning researcher points out how the field has become enshittified. Everything is about publications, beating benchmarks, and social media. LLM use in papers, LLM use in reviews, LLM use in meta-reviews. Nobody cares about the meaning of the actual research anymore.
lagrangeinterpolator
One of the few benefits of AI is that nowadays some PR threads are very entertaining to read.
“California is, I believe, the only state to give health insurance to people who come into the country illegally,” Kauffman said nervously. “I think we probably should not be providing that.”
“So you’d rather everyone just be sick, and get everyone else sick?” another reporter asked.
“That’s not what I’m saying,” said Kauffman.
“Isn’t that effectively what happens?” the reporter countered. “They don’t have access to health care and they just have to get sick, right?”
Kauffman contemplated that one for a moment. “Then they have to just get sick,” he said. “I mean, it’s unfortunate, but I think that it’s sort of impossible to have both liberal immigration laws and generous government benefits.”
Do I need to comment on this one?
I don't even think many AI developers realize that we're in a hype bubble. From what I see, they genuinely believe that the Models Will Improve and that These Issues Will Get Fixed. (I see a lot of faculty in my department who still have these beliefs.)
What these people do see, however, are a lot of haters who just cannot accept this wonderful new technology for some reason. AI is so magical that they don't need to listen to the criticisms; surely they're trivial by comparison to magic, and whatever they are, These Issues Will Get Fixed. But lately they have realized that with the constant embarrassing AI failures (AI surely doesn't have horrible ethics as well), there are a lot of haters who will swarm the announcement of any AI project now. The haters also tend to be people who actually know stuff and check things (tech journalists are incentivized to not do that), but that is an advantage because they're just random internet commenters, not big news outlets.
My theory is that now they add a ton of caveats and disclaimers to their announcements in a vain attempt to reduce the backlash. Also if you criticize them, it's actually your fault that it doesn't work. It's Still Early Days. These Issues Will Get Fixed.
I knew the Anthropic blog post was bullshit but every single time the reality is 10x worse that I anticipated.
I wonder what actual experts in compilers think of this. There were some similar claims about vibe coding a browser from scratch that turned out to be a little overheated: https://pivot-to-ai.com/2026/01/27/cursor-lies-about-vibe-coding-a-web-browser-with-ai/
I do not believe that this demonstrates anything other than they kept making the AI brute force random shit until it happened to pass all the test cases. The only innovation was that they spent even more money than before. Also, it certainly doesn't help that GCC is open source, and they have almost certainly trained the model on the GCC source code (which the model can regurgitate poorly into Rust). Hell, even their blog post talks about how half their shit doesn't work and just calls GCC instead!
It lacks the 16-bit x86 compiler that is necessary to boot Linux out of real mode. For this, it calls out to GCC (the x86_32 and x86_64 compilers are its own).
It does not have its own assembler and linker; these are the very last bits that Claude started automating and are still somewhat buggy. The demo video was produced with a GCC assembler and linker.
I wonder why this blog post was brazen enough to talk about these problems. Perhaps by throwing in a little humility, they can make the hype pill that much easier to swallow.
Sidenote: Rust seems to be the language of choice for a lot of these vibe coded "projects", perhaps because they don't want people immediately accusing them of plagiarism. But Rust syntax still reasonably follows languages like C. In most cases, blindly translating C code into Rust kinda works. Now, Rust does have the borrow checker which requires a lot of thinking to deal with, but I think this is not actually a disadvantage for the AI. Borrow checking is enforced by the compiler, so if you screw up in that department, your code won't even compile. This is great for an AI that is just brute forcing random shit until it "works".
I guess I can check back in six months to see how they're doing ... wait a minute, they were saying the same things six months ago, weren't they? That's a bummer.
$1000 a week?? Even putting aside literally all of the other issues of AI, it is quite damning that AI cannot even beat humans on cost. AI somehow manages to screw up the one undeniable advantage of software. How do these people delude themselves into thinking that the dogshit they're eating is good?
As a sidenote, I think after the bubble collapses, the people who predict that there will still be some uses for genAI are mostly wrong. In large part, this is because they do not realize just how ruinously expensive it is to run these models, let alone scrape data and train them. Right now, these costs are being subsidized by venture capitalists putting their money into a furnace.
I admire how persistent the AI folks are at failing to do the same thing over and over again, but each time coming up with an even more stupid name. Vibe coding? Gas Town? Clawdbot, I mean Moltbook, I mean OpenClaw? It's probably gonna be something different tomorrow, isn't it?
It's a big club and you ain't in it!
It took a full eleven paragraphs before the article even mentions AI. Before that, it was a bunch of stuff about how Wikipedia is conservative and Gen Z and Gen Alpha have no attention span. If the author has to bury the real point and attempt to force this particular rhetorical framing, I think the haters are winning. Well done everyone.
my comments about this turd of an article
Surely, AI summaries are exactly the same as stuff like VisualEditor and Media Viewer, which were tools that helped contributors improve articles. Please ignore my rhetorical sleight of hand. They're exactly the same! Okay, I did mention AI hallucinations in one sentence, but let's move on from that real quick.
Yeah, because Wikipedia editors are permanently static. Back in 2001, Jimmy Wales handpicked a bunch of teenagers to have the sacred title of Wikipedia Editor, and they are the only ones who will ever be allowed to edit Wikipedia. Oh wait, it doesn't work like that. Older people retire and move on, and new people join all the time.
Now that we have all these golden eggs, who needs the goose anymore? Actually, it is Inevitable that the goose must be killed. It is progress. It is the advancement of technology. We just have to accept it.
So AI is a parasite that takes from Wikipedia, contributes nothing in return, and in fact actively chokes it out? And you think the solution is for Wikipedia to just surrender and implement AI features? Do you keep forgetting what point you're trying to make?
Yeah, what a wonderful suggestion. The AI companies just never realized all this time that they could use legitimate channels and give back to the sources they use. It's not like they are choosing to do this because they have no ethics and want the number to go up no matter the costs to themselves or to others.
Wikipedia has survived countless predictions of its demise, but I'm sure this prediction of its demise is going to pan out. After all, AI is more important than electricity, probably.