Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Dwarkesh Patel

👤 Person
12212 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Hmm.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

What is a solution to model collapse?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

I mean, there's very naive things you could attempt.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

It's just like the distribution over logits should be wider or something.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Like, there's many naive things you could try.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

What ends up being the problem with the naive approaches?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

In fact, it's actively penalized, right?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

If you're like super creative in RL, it's like not good.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

And then I think you hinted that it's a very fundamental problem.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

It won't be easy to solve.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

What's your intuition for that?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

How many bits should the optimal core of intelligence end up being if you just had to make a guess?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

The thing we put on the von Neumann probes, how big does it have to be?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

That's actually surprising that you think it will take a billion, because already we have a billion parameter models, or a couple billion parameter models that are like very intelligent.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Well, some of our models are like a trillion parameters, right?

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

But they remember so much stuff.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Yeah, but I'm surprised that in 10 years, given the pace, okay, we have GPT-OSS-20b, that's way better than GPT-4 original, which was a trillion plus parameters.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Yeah.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

So given that trend, I'm actually surprised you think in 10 years, the cognitive core is still a billion parameters.

Dwarkesh Podcast
Andrej Karpathy — AGI is still a decade away

Yeah, I'm surprised you're not like, oh, it's going to be like tens of millions or millions.