Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Terence Tao

๐Ÿ‘ค Person
2047 total appearances

Appearances Over Time

Podcast Appearances

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

Just the aggregate behavior is important.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

But if you want to model a fluid, like the weather, you can't just say in Los Angeles, the temperature is this, the wind speed is this.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

For supercritical equations, the finite scale information is really important.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

Can you describe this idea?

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

Right, yeah.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

So this came out of this work of constructing this average equation that blew up.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

So one...

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

As part of how I had to do this, there's sort of this naive way to do it.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

You just keep pushing.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

Every time you get energy at one scale, you push it immediately to the next scale as fast as possible.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

This is sort of the naive way to force blow up.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

It turns out in five and higher dimensions, this works.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

But in three dimensions, there was this funny phenomenon that I discovered, that if you keep...

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

change the laws of physics, you just always keep trying to push the energy into smaller and smaller scales.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

What happens is that the energy starts getting spread out into many scales at once.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

So you have energy at one scale, you're pushing it into the next scale, and then

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

as soon as it enters that scale, you also push it to the next scale, but there's still some energy left over from the previous scale.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

You're trying to do everything at once.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

And this spreads out the energy too much.

Lex Fridman Podcast
#472 โ€“ Terence Tao: Hardest Problems in Mathematics, Physics & the Future of AI

And then it turns out that it makes it vulnerable for viscosity to come in and actually just damp out everything.