Folks in the field of AI like to make predictions for AGI. I have thoughts, and Iโve always wanted to write them down. Letโs do that.
Since this isnโt something Iโve touched on in the past, Iโll start by doing my best to define what I mean by โgeneral intelligenceโ: a generally intelligent entity is one that achieves a special synthesis of three things:
A way of interacting with and observing a complex environment. Typically this means embodiment: the ability to perceive and interact with the natural world.
A robust world model covering the environment. This is the mechanism which allows an entity to perform quick inference with a reasonable accuracy. World models in humans are generally referred to as โintuitionโ, โfast thinkingโ or โsystem 1 thinkingโ.
A mechanism for performing deep introspection on arbitrary topics. This is thought of in many different ways โ it is โreasoningโ, โslow thinkingโ or โsystem 2 thinkingโ.
If you have these three things, you can build a generally intelligent agent. Hereโs how:
First, you seed your agent with one or more objectives. Have the agent use system 2 thinking in conjunction with its world model to start ideating ways to optimize for its objectives. It picks the best idea and builds a plan. It uses this plan to take an action on the world. It observes the result of this action and compares that result with the expectation it had based on its world model. It might update its world model here with the new knowledge gained. It uses system 2 thinking to make alterations to the plan (or idea). Rinse and repeat.
My definition for general intelligence is an agent that can coherently execute the above cycle repeatedly over long periods of time, thereby being able to attempt to optimize any objective.
The capacity to actually achieve arbitrary objectives is not a requirement. Some objectives are simply too hard. Adaptability and coherence are the key: can the agent use what it knows to synthesize a plan, and is it able to continuously act towards a single objective over long time periods.
So with that out of the way โ where do I think we are on the path to building a general intelligence?
World Models
Weโre already building world models with autoregressive transformers, particularly of the โomnimodelโ variety. How robust they are is up for debate. Thereโs good news, though: in my experience, scale improves robustness and humanity is currently pouring capital into scaling autoregressive models. So we can expect robustness to improve.
With that said, I suspect the world models we have right now are sufficient to build a generally intelligent agent.
Side note: I also suspect that robustness can be further improved via the interaction of system 2 thinking and observing the real world. This is a paradigm we havenโt really seen in AI yet, but happens all the time in living things. Itโs a very important mechanism for improving robustness.
When LLM skeptics like Yann say we havenโt yet achieved the intelligence of a cat โ this is the point that they are missing. Yes, LLMs still lack some basic knowledge that every cat has, but they could learn that knowledge โ given the ability to self-improve in this way. And such self-improvement is doable with transformers and the right ingredients.
Reasoning
There is not a well known way to achieve system 2 thinking, but I am quite confident that it is possible within the transformer paradigm with the technology and compute we have available to us right now. I estimate that we are 2-3 years away from building a mechanism for system 2 thinking which is sufficiently good for the cycle I described above.
Embodiment
Embodiment is something weโre still figuring out with AI but which is something I am once again quite optimistic about near-term advancements. There is a convergence currently happening between the field of robotics and LLMs that is hard to ignore.
Robots are becoming extremely capable โ able to respond to very abstract commands like โmove forwardโ, โget upโ, โkick ballโ, โreach for objectโ, etc. For example, see what Figure is up to or the recently released Unitree H1.
On the opposite end of the spectrum, large Omnimodels give us a way to map arbitrary sensory inputs into commands which can be sent to these sophisticated robotics systems.
Iโve been spending a lot of time lately walking around outside talking to GPT-4o while letting it observe the world through my smartphone camera. I like asking it questions to test its knowledge of the physical world. Itโs far from perfect, but it is surprisingly capable. Weโre close to being able to deploy systems which can commit coherent strings of actions on the environment and observe (and understand) the results. I suspect weโre going to see some really impressive progress in the next 1-2 years here.
This is the field of AI I am personally most excited in, and I plan to spend most of my time working on this over the coming years.
TL;DR
In summary โ weโve basically solved building world models, have 2-3 years on system 2 thinking, and 1-2 years on embodiment. The latter two can be done concurrently. Once all of the ingredients have been built, we need to integrate them together and build the cycling algorithm I described above. Iโd give that another 1-2 years.
So my current estimate is 3-5 years for AGI. Iโm leaning towards 3 for something that looks an awful lot like a generally intelligent, embodied agent (which I would personally call an AGI). Then a few more years to refine it to the point that we can convince the Gary Marcusโ of the world.
Really excited to see how this ages. ๐
I can't believe they fucking got me with this one. I remember back in August(?) Epoch was getting quotes from top mathematicians like Tarrence Tao to review the benchmark and he was quoted saying like it would be a big deal for a model to do well on this benchmark, it will be several years before a model can solve all these questions organically etc so when O3 dropped and got a big jump from SotA, people (myself) were blown away. At the same time red flags were going up in my mind: Epoch was yapping about how this test was completely confidential and no one would get to see their very special test so the answers wouldn't get leaked. But then how in the hell did they evaluate this model on the test? There's no way O3 was run locally by Epoch at ~$1000 a question -> OAI had to be given the benchmark to run against in house -> maybe they had multiple attempts against it and were tuning the model/ recovering questions from api logs/paying mathematicians in house to produce answers to the problems so they could generate their own solution set??
No. The answer is much stupider. The entire company of Epoch ARE mathematicians working for OAI to make marketing grift to pump the latest toy. They got me lads, I drank the snake oil prepared specifically for people like me to drink :(