Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Tamay Besiroglu

๐Ÿ‘ค Person
878 total appearances

Appearances Over Time

Podcast Appearances

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

If you scale up data collection as well, I think it gets even stronger, like real-world data collection by deployment and so on.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

But building Shenzhen in a desert, that's a pretty... Like, if you think about the...

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

Pipeline.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

So, so far we have relied โ€“ first of all, we're relying on the entire semiconductor supply chain.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

That industry depends on tons of inputs and materials and whatever.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

It gets from probably tons of random places in the world.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

And creating that infrastructure, like doubling or tripling, whatever, that infrastructure โ€“

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

like the entire thing, that's very hard work, right?

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

So probably you couldn't even do it even if you just have shens in a letter.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

Like that would be even more expensive than that.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

And on top of that, so far we have been drawing heavily on the fact that we have built up this huge stock of data

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

over the past 30 years or something on the internet.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

Like, imagine you were trying to train a state-of-the-art model, but you only have, like, 100 billion tokens, right, to train on.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

That would be very difficult.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

So, in a certain sense, our entire economy...

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

has produced this huge amount of data on the internet that we are now using to train the models.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

It's plausible that in the future when you need to get new competencies added to these systems, the most efficient way to do that will be to try to leverage similar kind of modalities of data, which will also require this like

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

You would want to deploy the systems broadly because that's going to give you more data.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

Maybe you can get where you want to be without that, but it would just be less efficient if you're starting from scratch compared to if you're collecting a lot of data.

Dwarkesh Podcast
AGI is Still 30 Years Away โ€” Ege Erdil & Tamay Besiroglu

I think this is actually a motivation for why labs want their LLMs to be deployed widely because sometimes when you talk to ChatGPT, it's going to give you two responses and it's going to say, well, which one was good?