Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Blog Pricing

Andrej Karpathy

๐Ÿ‘ค Speaker
3419 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

You can talk about the heat dissipation of that because your heat dissipation grows as the surface area, which is growing a square, but your heat creation or generation is growing as a cube.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And so I just feel like physicists have all the right cognitive tools to approach problem solving in the world.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So I think because of that training, I always try to find the first order terms or the second order terms of everything.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

When I'm observing a system or a thing, I have a tangle of a web of ideas or knowledge in my world, in my mind.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And I'm trying to find what is the thing that actually matters?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

What is the first order component?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

How can I simplify it?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

How can I have a simple thing that actually shows that thing, right?

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

It shows an action.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And then I can tack on the other terms.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Maybe an example from one of my repos that I think illustrates it well is called micrograd.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

I don't know if you're familiar with this.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So micrograd is 100 lines of code that shows backpropagation.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

You can create neural networks out of simple operations like plus and times, et cetera, Lego blocks of neural networks.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

And you build up a computational graph, and you do a forward pass and a backward pass to get the gradients.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Now, this is at the heart of all neural network learning.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So MicroGrad is a 100 lines of pre-interpretable Python code, and it can do forward and backward arbitrary neural networks, but not efficiently.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

So MicroGrad, these 100 lines of Python, are everything you need to understand how neural networks train.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Everything else is just efficiency.

Dwarkesh Podcast
Andrej Karpathy โ€” AGI is still a decade away

Everything else is efficiency.