Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Dwarkesh Patel

👤 Person
12212 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

The difference, I guess, is that evolution...

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

has to be titrated in the case of humans through three gigabytes of DNA.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And so that's very unlike the weights of a model.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

I mean, literally the weights of the model are a brain, which obviously is not encoded in the sperm and the egg, or does not exist in the sperm and the egg.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

So it has to be grown.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And also the information for every single synapse in the brain simply cannot exist in the three gigabytes that exist in the DNA.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Evolution seems closer to finding the algorithm

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

which then does the lifetime learning.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Now, maybe the lifetime learning is not analogous to RL, to your point.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Is that compatible with the thing you were saying, or would you disagree with that?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Just to steel man the other perspective, because after doing this in an interview and thinking about it a bit, he has an important point here.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Evolution does not give us the knowledge, really, right?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

It gives us the algorithm to find the knowledge.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And that seems different from pre-training.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

So if perhaps the perspective is that pre-training helps build the kind of entity which can learn better, it teaches meta-learning.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

and therefore it is similar to like finding an algorithm.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

But if it's like evolution gives us knowledge and pre-training gives us knowledge, that analogy seems to break down.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

There's so much interesting stuff there.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Okay, so let's start with in-context learning.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

This is an obvious point, but I think it's worth just like saying it explicitly and meditating on it.