Some thoughts on the Sutton interview
I have a much better understanding of Sutton’s perspective now. I wanted to reflect on it a bit.
(00:00:00) - The steelman
(00:02:42) - TLDR of my current thoughts
(00:03:22) - Imitation learning is continuous with and complementary to RL
(00:08:26) - Continual learning
(00:10:31) - Concluding thoughts
Get full access to Dwarkesh Podcast at www.dwarkesh.com/subscribe
Listen and follow along
Transcript
Boy, do you guys have a lot of thoughts about the Sudden interview?
I've been thinking about it myself, and I think I have a much better understanding now of Sutton's perspective than I did during the interview itself.
So I wanted to reflect on how I understand his worldview now.
And Richard, apologies if there's still any errors or misunderstandings.
It's been very productive to learn from your thoughts.
Okay, so here's my understanding of the steel man of Richard's position.
Obviously, he wrote the famous essay, The Bitter Lesson.
And what is this essay about?
Well, it's not saying that you just want to throw away as much compute as you possibly can.
The bitter lesson says that you want to come up with techniques which most effectively and scalably leverage compute.
Most of the compute that's spent on an LLM is used in running it during deployment.
And yet it's not learning anything during this entire period.
It's only learning during the special phase that we call training.
And so this is obviously not an effective use of compute.
And what's even worse is that this training period by itself is highly inefficient because these models are usually trained on the equivalent of tens of thousands of years of human experience.
And what's more, during this training phase, all of their learning is coming straight from human data.
Now, this is an obvious point in the case of pre-training data, but it's even kind of true for the RLVR that we do with these LLMs.
These RLM environments are human-furnished playgrounds to teach LLMs the specific skills that we have prescribed for them.
The agent is in no substantial way learning from organic and self-directed engagement with the world.
Having to learn only from human data, which is an inelastic and hard-to-skill resource, is not a scalable way to use compute.
Furthermore, what these LLMs learn from training is not a true world model, which would tell you how the environment changes in response to different actions that you take.
Rather, they are building a model of what a human would say next.
And this leads them to rely on human-derived concepts.
A way to think about this would be, suppose you trained an LLM on all the data up to the year 1900.
That LLM probably wouldn't be able to come up with relativity from scratch.
And maybe here's a more fundamental reason to think this whole paradigm will eventually be superseded.
LLMs aren't capable of learning on the job, so we'll need some new architecture to enable this kind of continual learning.
And once we do have this architecture, we won't need a special training phase.
The agents will just be able to learn on the fly, like all humans, and in fact, like all animals are able to do.
And this new new paradigm will render our current approach with LLMs and their special training phase that's super sample and efficient totally obsolete.
So that's my understanding of Rich's position.
My main difference with Rich is just that I don't think the concepts he's using to distinguish LLMs from true intelligence or animal intelligence are actually that mutually exclusive or dichotomous.
For example, I think imitation learning is continuous with and complementary to RL.
And relatedly, models of humans can give you a prior which facilitates learning quote-unquote true world models.
I also wouldn't be surprised if some future version of test time fine-tuning could replicate continual learning, given that we've already managed to accomplish this somewhat with in-context learning.
So let's start with my claim that imitation learning is continuous with and complementary to RL.
So I tried to ask Richard a couple of times whether pre-trained LLMs can serve as a good prior on which we can accumulate the experiential learning, aka do the RL, which would lead to AGI.
So Ilyas Escover gave a talk a couple of months ago that I thought was super interesting, and he compared pre-training data to fossil fuels.
And I think this analogy actually has remarkable reach.
Just because fossil fuels are not a renewable resource does not mean that our civilization ended up on a dead-end track by using them.
In fact, they were absolutely crucial.
You simply couldn't have transitioned from the water wheels of 1800 to solar panels and fusion power plants.
We had to use this cheap, convenient, and plentiful intermediary to get to the next step.
AlphaGo, which was conditioned on human games, and AlphaZero, which was bootstrapped from scratch, were both superhuman Go players.
Now, of course, AlphaZero was better.
So you can ask the question, will we or will the first AGIs eventually come up with a general learning technique that requires no initialization of knowledge and that just bootstraps itself from the very start?
And will it outperform outperform the very best AIs that have been trained up to that date?
I think the answer to both these questions is probably yes.
But does this mean that imitation learning must not play any role whatsoever in developing the first AGI or even the first ASI?
No.
AlphaGo is still superhuman despite being initially shepherded by human player data.
The human data isn't necessarily actively detrimental.
It's just that at enough scale, it isn't significantly helpful.
AlphaZero also used much more compute than AlphaGo.
The accumulation of knowledge over tens of thousands of years has clearly been essential to humanity's success.
In any field of knowledge, thousands and probably actually millions of previous people were involved in building up our understanding and passing it on to the next generation.
We obviously didn't invent the language we speak nor the legal system we use.
Also, even most of the technologies in your phone were not directly invented by the people who are alive today.
This process is more analogous to imitation learning than it is to RL from scratch.
Now, of course, are we literally predicting the next token like an LLM would in order to do this cultural learning?
No, of course not.
So even the imitation learning that humans are doing is not like the supervised learning that we do for pre-training LLMs.
But neither are we running around trying to collect some well-defined scalar reward.
No ML learning regime perfectly describes human learning or animal learning.
We're doing things which are both analogous to RL and to supervised learning.
What planes are to birds, supervised learning might end up being to human cultural learning.
I also don't think these learning techniques are actually categorically different.
Imitation learning is just short horizon RL.
The episode is a token long.
The LLM is making a conjecture about the next token based on its understanding of the world and how the different pieces of information in the sequence relate to each other.
And it receives reward in proportion to how well it predicted the next token.
Now, of course, I already hear people saying, no, no, that's not the ground truth.
It's just learning what a human was likely to say.
And I agree, but there's a different question, which I think is actually more relevant to understanding the scalability of these models.
And that question is, can we leverage this imitation learning to help models learn better from ground truth?
And I think the answer is obviously yes.
After RLing these pre-trained base models, we've gotten them to win gold in international Math Olympiad competitions and to code up entire working applications from scratch.
Now, these are ground ground truth examinations.
Can you solve this unseen Math Olympiad question?
Can you build this application to match a specific feature request?
But you couldn't have RL'd a model to accomplish these tasks from scratch, or at least we don't know how to do that yet.
You needed a reasonable prior over human data in order to kickstart this RL process.
Whether you want to call this prior a proper world model or just a model of humans, I don't think is that important.
It honestly seems like a semantic debate.
Because what you really care about is whether this model of humans helps you start learning from ground truth, aka become a true world model.
It's a bit like saying to somebody pasteurizing milk, hey, you should stop boiling that milk because eventually you want to serve it cold.
Of course,
but this is an intermediate step to facilitate the final output.
By the way, LLMs are clearly developing a deep representation of the world because their training process is incentivizing them to develop one.
I use LLMs to teach me about everything from biology to AI to history, and they are able to do so with remarkable flexibility and coherence.
Now, are LLMs specifically trained to model how their actions will affect the world?
No, they are not.
But if we're not allowed to call their representations a world model, then we're defining the term world model by the process that we think is necessary to build one rather than the obvious capabilities that this concept implies.
Okay, continual learning.
I'm sorry to bring up my hobby horse again.
I'm like a comedian who has only come up with one good bit, but I'm going to milk it for all it's worth.
An LLM that's being RL'd on outcome-based rewards learns on the order of one bit per episode.
And an episode might be tens of thousands of tokens long.
Now, obviously, animals and humans are clearly extracting more information from interacting with our environment than just the reward signal at the end of an episode.
Conceptually, how should we think about what is happening with animals?
I think we're learning to model the world through observations.
This outer loop RL is incentivizing some other learning system to pick up maximum signal from the environment.
In Richard's Oak Architecture, he calls this the transition model.
And if we were trying to pigeonhole this feature spec into modern LLMs, what you do is fine-tune on all your observed tokens.
From what I hear from my researcher friends, in practice, the most naive way of doing this actually doesn't work very well.
Now, being able to learn from the environment in a high-throughput way is obviously necessary for true HEI, and it clearly doesn't exist with LLMs trained on RLVR.
But there might be some other relatively straightforward ways to shoehorn continual learning atop LLMs.
For example, one could imagine making supervised fine-tuning a tool call for the model.
So the outer loop RL is incentivizing the model to teach itself effectively using supervised learning in order to solve problems that don't fit in the context window.
Now, I'm genuinely agnostic about how well techniques like this will work.
I'm not an AI researcher, but I wouldn't be surprised if they basically replicate continual learning.
And the reason is that models are already demonstrating something resembling human continual learning within their context windows.
The fact that in-context learning emerged spontaneously from the training incentive to process long sequences makes me think that if information could just flow across windows longer than the context limit, then models could meta-learn the same flexibility that they already show in context.
Okay, some concluding thoughts.
Evolution does meta RL to make an RL agent, and that agent can selectively do imitation learning.
With LLMs, we're going the opposite way.
We have first made this base model that does pure imitation learning, and then we're hoping that we do enough RL on it to make a coherent agent with goals and self-awareness.
Maybe this won't work.
But I don't think these super first principles arguments about, for example, how these LMs don't have a true world model are actually proving much.
And I I also don't think they're strictly accurate for the models we have today, which are actually undergoing a lot of RL on ground truth.
Even if Sun's platonic ideal doesn't end up being the path to the first AGI, his first principles critique is identifying some genuine basic gaps that these models have.
And we don't even notice them because they're so pervasive in the current paradigm, but because he has this decades-long perspective, they're obvious to him.
It's the lack of continual learning, it's the abysmal sample efficiency of these models, it's their dependence on exhaustible human data.
If the LLMs do get to HEI first, which is what I expect to happen, the successor systems that they build will almost certainly be based on Richard's vision.