Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Richard Sutton

👤 Person
505 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

We're not seeing general... Critical to good performance is that you can generalize well from one state to another state.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

We don't have any methods that are good at that.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

What we have are people try different things and they settle on something that a representation that transfers well or that generalizes well.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

But we don't have any automated techniques to promote.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

We have very few automated techniques to promote transfer.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

And none of them are used in modern deep learning.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

The researchers did it.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

Because there's no other explanation.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

Gradient descent will not make you generalize well.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

It will make you solve the problem.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

It will not make you get new data.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

you generalize in a good way.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

Generalization means train on one thing that affects what you do on the other things.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

So we know deep learning is really bad at this.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

For example, we know that if you train on some new thing, it will often catastrophically interfere with all the old things that you knew.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

So this is exactly bad generalization.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

Right.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

Generalization, as I said, is some kind of

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

influence of training on one state on other states.

Dwarkesh Podcast
Richard Sutton – Father of RL thinks LLMs are a dead-end

And generalization is not necessarily good or bad.