lagrangeinterpolator

joined 9 months ago
[–] lagrangeinterpolator@awful.systems 2 points 4 hours ago* (last edited 4 hours ago)

Hey, you're selling them short: there are also ReLU and softmax activation functions thrown around here and there. Clankers aren't just linear transformations! /j

[–] lagrangeinterpolator@awful.systems 5 points 5 hours ago* (last edited 5 hours ago) (1 children)

I am a computer science PhD so I can give some opinion on exactly what is being solved.

First of all, the problem is very contrived. I cannot think of what the motivation or significance of this problem is, and Knuth literally says that it is a planned homework exercise. It's not a problem that many people have thought about before.

Second, I think this problem is easy (by research standards). The problem is of the form: "Within this object X of size m, find any example of Y." The problem is very limited (the only thing that varies is how large m is), and you only need to find one example of Y for each m, even if there are many such examples. In fact, Filip found that for small values of m, there were tons of examples for Y. In this scenario, my strategy would be "random bullshit go": there are likely so many ways to solve the problem that a good idea is literally just trying stuff and seeing what sticks. Knuth did say the problem was open for several weeks, but:

  1. Several weeks is a very short time in research.
  2. Only he and a couple friends knew about the problem. It was not some major problem many people were thinking about.
  3. It's very unlikely that Knuth was continuously thinking about the problem during those weeks. He most likely had other things to do.
  4. Even if he was thinking about it the whole time, he could have gotten stuck in a rut. It happens to everyone, no matter how much red site/orange site users worship him for being ultra-smart.

I guess "random bullshit go" is served well by a random bullshit machine, but you still need an expert who actually understands the problem to read the tea leaves and evaluate if you got something useful. Knuth's narrative is not very transparent about how much Filip handheld for the AI as well.

I think the main danger of this (putting aside the severe societal costs of AI) is not that doing this is faster or slower than just thinking through the problem yourself. It's that relying on AI atrophies your ability to think, and eventually even your ability to guard against the AI bullshitting you. The only way to retain a deep understanding is to constantly be in the weeds thinking things through. We've seen this story play out in software before.

I was pissed when my (non-academic) friends saw this and immediately started talking about how mathematicians and computer scientists need to use AI from now on.

[–] lagrangeinterpolator@awful.systems 9 points 8 hours ago* (last edited 8 hours ago)

Baldur Bjarnason's essay remains evergreen.

Consider homeopathy. You might hear a friend talk about “water memory”, citing all sorts of scientific-sounding evidence. So, the next time you have a cold you try it.

And you feel better. It even feels like you got better faster, although you can’t prove it because you generally don’t document these things down to the hour.

“Maybe there is something to it.”

Something seemingly working is not evidence of it working.

  • Were you doing something else at the time which might have helped your body fight the cold?

  • Would your recovery have been any different had you not taken the homeopathic “remedy”?

  • Did your choosing of homeopathy over established medicine expose you to risks you weren’t aware of?

Even when looking at Knuth's account of what happened, you can already tell that the AI is receiving far more credit than what it actually did. There is something about a nondeterministic slot machine that makes it feel far more miraculous when it succeeds, while reliable tools that always do their job are boring and stupid. The downsides of the slot machine never register in comparison to the rewards.

I feel like math research is particularly susceptible to this, because it is the default that almost all of one's attempts do not succeed. So what if most of the AI's attempts do not succeed? But if it is to be evaluated as a tool, we have to check if the benefits outweigh the costs. Did it give me more productive ideas, or did it actually waste more of my time leading me down blind alleys? More importantly, is the cognitive decline caused by relying on slot machines going to destroy my progress in the long term? I don't think anyone is going to do proper experiments for this in math research, but we have already seen this story play out in software. So many people were impressed by superficial performances, and now we are seeing the dumpster fire of bloat, bugs, and security holes. No, I don't think I want that.

And then there is the narrative of not evaluating AI as an objective tool based on what it can actually do, but instead as a tidal wave of Unending Progress that will one day sweep away those elitists with actual skills. This is where the AI hype comes from, and why people avoid, say, comparing AI with Mathematica. To them I say good luck. We have dumped hundreds of billions of dollars into this, and there are only so many more hundreds of billions of dollars left. Were these small positive results (and significant negatives in places like software) worth hundreds of billions of dollars, or perhaps were there better things that these resources could have been used for?

Don't worry, there's always Effective Altruism if you ever feel guilty about causing the suffering of regular people. Just say you're going to donate your money at some point eventually in the future. There you go, 40 trillion hypothetical lives saved!

[–] lagrangeinterpolator@awful.systems 7 points 1 week ago* (last edited 1 week ago) (1 children)

This somehow makes things even funnier. If he had any understanding of modern math, he would know that representing a set of things as points in some geometric space is one of the most common techniques in math. (A basic example: a pair of numbers can be represented by a point in 2D space.) Also, a manifold is an extremely broad geometric concept: knowing that two things are manifolds does not meant that they are the same or even remotely similar, without checking the details. There are tons of things you can model as a manifold if you try hard enough.

From what I see, Scoot read a paper modeling LLM inference with manifolds and thought "wow, cool!" Then he fished for neuroscience papers until he found one that modeled neurons using manifolds. Both of the papers have blah blah blah something something manifolds so there must be a deep connection!

(Maybe there is a deep connection! But the burden of proof is on him, and he needs to do a little more work than noticing that both papers use the word manifold.)

[–] lagrangeinterpolator@awful.systems 3 points 1 week ago (1 children)

Kolmogorov complexity:

So we should see some proper definitions and basic results on the Kolmogorov complexity, like in modern papers, right? We should at least see a Kt or a pKt thrown in there, right?

Understanding IS compression — extracting structure from data. Optimal compression is uncomputable. Understanding is therefore always provisional, always improvable, never verifiably complete. This kills “stochastic parrot” from a second independent direction: if LLMs were memorizing rather than understanding, they could not generalize to inputs not in their training data. But they do. Generalization to novel input IS compression — extracting structure, not regurgitating sequences.

Fuck!

[–] lagrangeinterpolator@awful.systems 12 points 1 week ago (8 children)

Nonsensical analogies are always improved by adding a chart with colorful boxes and arrows going between them. Of course, the burden of proof is on you, dear reader, to explain why the analogy doesn't make sense, not on the author to provide more justification than waving his hands really really hard.

Many of these analogies are bad as, I don't know, "Denmark and North Korea are the same because they both have governments" or something. Humans and LLMs both produce sequences of words, where the next word depends in some way on the previous words, so they are basically the same (and you can call this "predicting" the next word as a rhetorical flourish). Yeah, what a revolutionary concept, knowing that both humans and LLMs follow the laws of time and causality. And as we know, evolution "optimizes" for reproduction, and that's why there are only bacteria around (they can reproduce every 20 minutes). He has to be careful, these types of dumbass "optimization" interpretations of evolution that arose in the late 1800s led to horrible ideas about race science ... wait a minute ...

He isn't even trying with the yellow and orange boxes. What the fuck do "high-D toroidal attractor manifolds" and "6D helical manifolds" have to do with anything? Why are they there? And he really thinks he can get away with nobody closely reading his charts, with the "(???, nothing)" business. Maybe I should throw in that box in my publications and see how that goes.

I feel like his arguments rely on the Barnum effect. He makes statements like "humans and LLMs predict the next word" and "evolution optimizes for reproduction" that are so vague that they can be assigned whatever meaning he wants. Because of this, you can't easily dispel them (he just comes up with some different interpretation), and he can use them as carte blanche to justify whatever he wants.

[–] lagrangeinterpolator@awful.systems 11 points 1 week ago* (last edited 1 week ago)

Maybe I should apply to be a director of AI safety at Meta. I know one safety measure that works: don't use AI.

[–] lagrangeinterpolator@awful.systems 8 points 1 week ago (1 children)

What's next, are the crypto bros gonna make some dumb talking point about how traditional finance also uses so much energy ... oh wait, they already did that.

[–] lagrangeinterpolator@awful.systems 9 points 1 week ago* (last edited 1 week ago) (3 children)

For all the talk about these people being "highly agentic", it is deeply ironic how all the shit they do has no meaning and purpose. I hear all this sound and fury about making millions off of ChatGPT wrappers, meeting senators in high school bathrooms, and sperm races (?), and I wonder what the point is. Silicon Valley hagiographies used to at least have a veneer that all of this was meaningful. Are we supposed to emulate anyone just because they happen to temporarily have a few million dollars?

Even though the material conditions of working in science are not good, I'd still rather do science than whatever the hell they're doing. I would be sick at the prospect of being a "highly agentic" person in a "new and possibly permanent overclass", where my only sense of direction is a vague voice in my head telling me that I should be optimizing my life in various random ways, and my only motivation is the belief that I have to win harder and score more points on the leaderboard. (In any case, I believe this "overclass" is a lot more fragile than the author seems to think.)

view more: next ›