Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrej Karpathy

๐Ÿ‘ค Speaker
3419 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's some like stock ticker symbols.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's a huge amount of slop and garbage from like all the corners of the internet.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's not like your Wall Street Journal article that's extremely rare.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So I almost feel like because the internet is so terrible, we actually have to sort of like build really big models to compress all that.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Most of that compression is memory work instead of like cognitive work.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

But what we really want is the cognitive part to actually delete the memory.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And then, so I guess what I'm saying is like we need

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

intelligent models to help us refine even the pre-training set to just narrow it down to the cognitive components.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And then I think you get away with a much smaller model because it's a much better data set and you could train it on it.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

But probably it's not trained directly on it.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It's probably distilled for a much better model still.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I just feel like distillation works extremely well.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So almost every small model, if you have a small model, it's almost certainly distilled.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I mean, come on, right?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I don't know.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

At some point, it should take at least a billion knobs to do something interesting.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

You're thinking it should be even smaller?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I mean, I almost feel like I'm already contrarian by talking about a billion-parameter cognitive core, and you're outdoing me.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I think, yeah, maybe we could get a little bit smaller.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I mean, I still think that there should be enough.