Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrej Karpathy

๐Ÿ‘ค Speaker
3433 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Yeah.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Or like maybe if you're doing a lot of writing help from LLMs and stuff like that, I think it's probably bad because the models will give you these like silently all the same stuff, you know.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So they're not, they won't explore lots of different ways of answering a question, right?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

But I kind of feel like maybe this diversity is just not as big of a, yeah, maybe like, yeah, not as many applications needed so the models don't have it, but then it's actually a problem with synthetic generation time, et cetera.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So we're actually shooting ourselves in the foot by not allowing this entropy to maintain in the model.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And I think possibly the labs should try harder.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Yeah.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I don't actually know if it's super fundamental.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I don't actually know if I intended to say that.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I do think that...

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I haven't done these experiments, but I do think that you could probably regularize the entropy to be higher.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So you're encouraging the model to give you more and more solutions.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

But you don't want it to start deviating too much from the training data.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's going to start making up its own language.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's going to start using words that are extremely rare.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So it's going to drift too much from the distribution.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So I think controlling the distribution is just like a tricky... It's just like someone just has to...

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's probably not trivial in that sense.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So it's really interesting in the history of the field because at one point everything was very scaling-pilled in terms of like, oh, we're going to make much bigger models, trillions of parameter models.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And actually what the models have done in size is they've gone up and now they've actually kind of like