Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing

Aman Sanger

👤 Person
1050 total appearances

Appearances Over Time

Podcast Appearances

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

I think test time compute is really, really interesting. So there's been the pre-training regime, which will kind of, as you scale up the amount of data and the size of your model, get you better and better performance, both on loss and then on downstream benchmarks and just general performance when we use it for coding or other tasks.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

We're starting to hit a bit of a data wall, meaning it's going to be hard to continue scaling up this regime.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

We're starting to hit a bit of a data wall, meaning it's going to be hard to continue scaling up this regime.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

We're starting to hit a bit of a data wall, meaning it's going to be hard to continue scaling up this regime.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

And so scaling up test time compute is an interesting way of now, you know, increasing the number of inference time flops that we use, but still getting like, like, yeah, as you increase the number of flops use inference time getting corresponding improvements in the performance of these models tremendously.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

And so scaling up test time compute is an interesting way of now, you know, increasing the number of inference time flops that we use, but still getting like, like, yeah, as you increase the number of flops use inference time getting corresponding improvements in the performance of these models tremendously.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

And so scaling up test time compute is an interesting way of now, you know, increasing the number of inference time flops that we use, but still getting like, like, yeah, as you increase the number of flops use inference time getting corresponding improvements in the performance of these models tremendously.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

Traditionally, we just had to literally train a bigger model that always used that many more flops. But now we could perhaps use the same size model and run it for longer to be able to get an answer at the quality of a much larger model. And so the really interesting thing I like about this is there are some problems that perhaps require

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

Traditionally, we just had to literally train a bigger model that always used that many more flops. But now we could perhaps use the same size model and run it for longer to be able to get an answer at the quality of a much larger model. And so the really interesting thing I like about this is there are some problems that perhaps require

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

Traditionally, we just had to literally train a bigger model that always used that many more flops. But now we could perhaps use the same size model and run it for longer to be able to get an answer at the quality of a much larger model. And so the really interesting thing I like about this is there are some problems that perhaps require

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

hundred trillion parameter model intelligence trained on a hundred trillion tokens. Um, but that's like maybe 1%, maybe like 0.1% of all queries. So are you going to spend all of this effort, all of this compute training model, uh,

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

hundred trillion parameter model intelligence trained on a hundred trillion tokens. Um, but that's like maybe 1%, maybe like 0.1% of all queries. So are you going to spend all of this effort, all of this compute training model, uh,

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

hundred trillion parameter model intelligence trained on a hundred trillion tokens. Um, but that's like maybe 1%, maybe like 0.1% of all queries. So are you going to spend all of this effort, all of this compute training model, uh,

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

that costs that much and then run it so infrequently, it feels completely wasteful when instead you get the model that can, that you train the model that's capable of doing the 99.9% of queries, then you have a way of inference time running it longer for those few people that really, really want max intelligence.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

that costs that much and then run it so infrequently, it feels completely wasteful when instead you get the model that can, that you train the model that's capable of doing the 99.9% of queries, then you have a way of inference time running it longer for those few people that really, really want max intelligence.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

that costs that much and then run it so infrequently, it feels completely wasteful when instead you get the model that can, that you train the model that's capable of doing the 99.9% of queries, then you have a way of inference time running it longer for those few people that really, really want max intelligence.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

I mean, yeah, that's an open research problem, certainly. I don't think anyone's actually cracked this model routing problem quite well. We'd like to. We have initial implementations of this for something like CursorTab. But at the level of going between 4.0 Sonnet to O1, It's a bit trickier.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

I mean, yeah, that's an open research problem, certainly. I don't think anyone's actually cracked this model routing problem quite well. We'd like to. We have initial implementations of this for something like CursorTab. But at the level of going between 4.0 Sonnet to O1, It's a bit trickier.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

I mean, yeah, that's an open research problem, certainly. I don't think anyone's actually cracked this model routing problem quite well. We'd like to. We have initial implementations of this for something like CursorTab. But at the level of going between 4.0 Sonnet to O1, It's a bit trickier.

Lex Fridman Podcast
#447 – Cursor Team: Future of Programming with AI

There's also a question of what level of intelligence do you need to determine if the thing is too hard for the four-level model? Maybe you need the O1-level model. It's really unclear.