Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Ilya Sutskever

👤 Person
766 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And so you say, okay, so what are we scaling?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And pre-training was a thing to scale.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

It was a particular scaling recipe.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

The big breakthrough of pre-training is

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

is the realization that this recipe is good so you say hey if you mix some compute with some data into a neural net of a certain size you will get results and you will know that will be better if you just scale the recipe up

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And this is also great.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Companies love this because it gives you a very low-risk way of investing your resources.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

It's much harder to invest your resources in research.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Compare that.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

If you research, you need to have go-forth researchers and research and come up with something versus get more data, get more compute, you know you'll get something from pre-training.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And indeed, you know, it looks like based on various things, some people say on Twitter, maybe it appears that Gemini have found a way to get more out of pre-training.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

At some point, though, pre-training will run out of data.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

The data is very clearly finite.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And so then, okay, what do you do next?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Either you do some kind of souped up retraining, different recipe from the one you've done before, or you're doing RL, or maybe something else.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

But now that compute is big, compute is now very big.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

In some sense, we are back to the age of research.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

So maybe here's another way to put it.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Up until 2020, from 2012 to 2020, it was the age of research.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Now, from 2020 to 2025, it was the age of scaling.