Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Dwarkesh Patel

👤 Person
12212 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

I mean, obviously, the very basic analogy would just be fine-tuning on reflection bits, but I feel like in practice that probably wouldn't work that well.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

So I don't know if you have some take on what the analogy of this thing is.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Just to make sure I understood, the reason that the collapse is relevant to synthetic data generation is because you want to be able to come up with synthetic problems or reflections which are not already in your data distribution?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

I guess what I'm saying is...

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

You can't just keep scaling, quote-unquote, reflection on the same amount of prompt information and then get returns from that.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Have you seen this super interesting paper that dreaming is a way of preventing this kind of overfitting and collapse?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

That the reason dreaming is an evolutionary adaptive is to put you in weird situations that are very unlike your day-to-day reality to prevent this kind of overfitting?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

This is a very ill-formed thought, so I'll just put it out and let you react to it.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

The best learners that we are aware of, which are children, are extremely bad at recollecting information.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

In fact, at the very earliest stages of childhood, you will forget everything.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

You're just an amnesiac about everything that happens before a certain year date.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

But you're extremely good at picking up new languages and learning from the world.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And maybe there's some element of being able to see the forest for the trees.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Whereas if you compare it to the opposite end of the spectrum, you have...

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

LLM pre-training, which these models were literally able to regurgitate word for word what is the next thing in a Wikipedia page.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

But their ability to learn abstract concepts really quickly the way a child can is much more limited.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And then adults are somewhere in between where they don't have the flexibility of childhood learning, but they can, you know, adults can memorize facts and information in a way that is harder for kids.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And I don't know if there's something interesting about that

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And this is also relevant to preventing model collapse.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Let me think.