Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Dwarkesh

👤 Person
1735 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

What is the kind of relationship we should be seeking and how should we think about what this new recipe might look like?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

So if we're back in the era of research, stepping back, what is the part of the recipe that we need to think most about?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

When you say value function, people are already trying the current recipe, but then having LLM as a judge and so forth.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

You could say that's a value function, but it sounds like you have something much more fundamental in mind.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Do we need to go back to, should we even rethink pre-training at all and not just add more steps to the end of that process?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

Okay, so this is the crux, generalization.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And there's two sub-questions.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

There's one which is about sample efficiency, which is why should it take so much more data for these models to learn than humans?

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

There's a second about even separate from the amount of data it takes.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

There's a question of why is it so hard to teach the thing we want to a model than to a human, which is to say for to a human that we don't necessarily need a verifiable reward to be able to.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

you're probably mentoring a bunch of researchers right now and you're, you know, talking with them, you're showing them your code and you're showing them how you think.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And from that, they're picking up your way of thinking and how they should do research.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

You don't have to set like a verifiable reward for them.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

That's like, okay, this is the next part of your curriculum.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And now this is the next part of your curriculum.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

And oh, it was, this training was unstable and we got to, there's not this schleppy bespoke process.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

So perhaps these two issues are actually related in some way, but I'd be curious to,

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

explore this second thing, which feels more like continual learning, and this first thing, which feels just like sample efficiency.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

It still seems better than models.

Dwarkesh Podcast
Ilya Sutskever – We're moving from the age of scaling to the age of research

I mean, obviously models are better than the average human at language and math and coding, but are they better at the average human at learning?