Mark Zuckerberg
๐ค SpeakerAppearances Over Time
Podcast Appearances
So...
I think that that's quite good.
Then there's the reasoning phenomenon, which you basically are alluding to with talking about 03 and 04 and some of the other models.
I do think that there is this specialization that's happening where if you want to model
that is sort of the best at math problems or coding or different things like that.
I do think that these reasoning models with a lot of the ability to just consume more test time or inference time compute in order to provide more intelligence is a really compelling paradigm.
But for a lot of the applications that, and we're gonna do that too, we're building a Lama 4 reasoning model and that'll come out at some point.
For a lot of the things that we care about,
Latency and good intelligence per cost are actually much more important product attributes.
If you're primarily designing for a consumer product, people don't necessarily want it to wait like half a minute to go think through the answer.
If you can provide an answer that's generally quite good too in like half a second, then that's great and that's a good trade-off.
So I think that both of these are gonna end up being
important directions.
I am optimistic about integrating the reasoning models with kind of the core language models over time.
I think that's sort of the direction that Google has gone in with some of the more recent Gemini models.
And I think that that's really promising.
But I think that there's just going to be a bunch of different stuff that goes on.
I mean, you also mentioned the whole chatbot arena thing, which I think is interesting.
And it goes to this challenge around how do you do the benchmarking, right?
And basically, how do you know what models are good for which things?